CN112232374B - Irrelevant label filtering method based on depth feature clustering and semantic measurement - Google Patents
Irrelevant label filtering method based on depth feature clustering and semantic measurement Download PDFInfo
- Publication number
- CN112232374B CN112232374B CN202010992837.4A CN202010992837A CN112232374B CN 112232374 B CN112232374 B CN 112232374B CN 202010992837 A CN202010992837 A CN 202010992837A CN 112232374 B CN112232374 B CN 112232374B
- Authority
- CN
- China
- Prior art keywords
- semantic
- cluster
- label
- label set
- labels
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
- G06F18/232—Non-hierarchical techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computational Linguistics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Evolutionary Biology (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
本发明公开了一种基于深度特征聚类和语义度量的不相关标签过滤方法,包括以下步骤:步骤一、传感器获取图像集;步骤二、建立与图像集对应的标签集;步骤三、提取图像集图像的深度特征;步骤四、对深度特征聚类获取聚类簇;步骤五、构建聚类簇的相关语义标签集合;步骤六、构建聚类簇的待度量标签集合;步骤七、生成语义向量;步骤八、计算语义向量的相关度;步骤九、根据相关度进行不相关标签过滤。本发明对庞大的样本图像数据聚类获取聚类簇,用于对样本图像数据的预分类,通过对聚类成的样本图像数据进行分析,拥有更高的有效性与正确性,同时对标签语义进行相关度度量,从而实现了不相关标签的自动过滤,可以提升深度网络的泛化性与鲁棒性。
The invention discloses a method for filtering irrelevant tags based on deep feature clustering and semantic measurement, comprising the following steps: step 1, a sensor acquires an image set; step 2, establishing a tag set corresponding to the image set; step 3, extracting an image Collect the depth features of the image; step 4, cluster the depth features to obtain clusters; step 5, construct the relevant semantic label set of the cluster; step 6, construct the set of labels to be measured for the cluster; step 7, generate semantics vector; Step 8, calculating the relevance of the semantic vector; Step 9, filtering irrelevant tags according to the relevance. The present invention clusters huge sample image data to obtain cluster clusters for pre-classification of sample image data. By analyzing the clustered sample image data, it has higher validity and correctness, and at the same time labels Semantics is used to measure the relevance, so as to realize the automatic filtering of irrelevant labels, which can improve the generalization and robustness of the deep network.
Description
技术领域technical field
本发明属于深度学习技术领域,具体涉及一种基于深度特征聚类和语义度量的不相关标签过滤方法。The invention belongs to the technical field of deep learning, and in particular relates to an irrelevant label filtering method based on deep feature clustering and semantic measurement.
背景技术Background technique
随着人工智能技术的发展,深度学习技术得到了广泛的应用,已经成为人们工作、生活中不可缺少的一部分,具体表现在计算机视觉和人工智能领域。深度学习技术是机器学习的分支,是一种以人工神经网络为架构,对数据进行表征学习的算法。With the development of artificial intelligence technology, deep learning technology has been widely used and has become an indispensable part of people's work and life, especially in the fields of computer vision and artificial intelligence. Deep learning technology is a branch of machine learning. It is an algorithm that uses artificial neural networks as the framework to perform representation learning on data.
YannLecun等人提出的卷积神经网络被广泛的成功应用于检测、分割、物体识别等各个图像领域。这些应用都使用了大量的有标签的数据。深度学习技术能够取得良好成果的前提是拥有海量训练数据,海量训练数据的获取就需要大量人员对这些训练数据进行标注,这一过程要付出昂贵的人力和物力成本。即使是利用无标签的数据结合无监督技术训练网络得到预训练模型,也是需要训练数据的语义分布与所要预测的数据间存在相关性才能得到泛化能力比较强的模型。The convolutional neural network proposed by YannLecun et al. has been widely and successfully applied to various image fields such as detection, segmentation, and object recognition. These applications all use large amounts of labeled data. The prerequisite for deep learning technology to achieve good results is to have massive training data. The acquisition of massive training data requires a large number of people to label these training data. This process requires expensive manpower and material costs. Even if the pre-training model is obtained by using unlabeled data combined with unsupervised technology to train the network, there is still a correlation between the semantic distribution of the training data and the data to be predicted in order to obtain a model with relatively strong generalization ability.
人工标注标签的过程繁琐复杂。针对深度学习不同任务,例如目标检测、语义分割等,由于数据来源的多样性,样本数据中会有一些样本信息和标签不相关,样本信息的关键词标签对样本的审核、检索和组织起到关键作用,因此标注的不相关性容易导致标注信息不能准确地反映样本数据的特征,并且会导致深度学习模型拟合参数的时间加长,效率较低,对于结构复杂的深度学习神经网络、层数多的深度学习神经网络来说尤其如此。数据的误标注问题一直是计算机视觉和人工智能的重点研究领域,因此为了提高深度学习模型的效率,需要研究数据集不相关标签过滤技术。The process of manual labeling is tedious and complicated. For different deep learning tasks, such as target detection, semantic segmentation, etc., due to the diversity of data sources, some sample information and labels in the sample data are irrelevant, and the keyword labels of sample information play an important role in the review, retrieval and organization of samples. The key role, so the irrelevance of annotations will easily lead to the fact that the annotation information cannot accurately reflect the characteristics of the sample data, and will lead to a longer time for the deep learning model to fit the parameters, and the efficiency is low. For the deep learning neural network with complex structure, the number of layers This is especially true for many deep learning neural networks. The problem of data mislabeling has always been a key research area of computer vision and artificial intelligence. Therefore, in order to improve the efficiency of deep learning models, it is necessary to study irrelevant label filtering technology for data sets.
目前的现有技术无法满足数据集不相关标签过滤要求,因此迫切需要一种对数据集中的不相关标签进行过滤,为后续的深度学习任务提供便利,可以提升深度网络的泛化性与鲁棒性的的数据集不相关标签过滤方法。The current existing technology cannot meet the requirements of filtering irrelevant labels in the data set, so there is an urgent need to filter irrelevant labels in the data set to facilitate subsequent deep learning tasks and improve the generalization and robustness of deep networks. A method for filtering irrelevant labels in datasets.
发明内容Contents of the invention
本发明所要解决的技术问题在于针对上述现有技术中的不足,提供一种基于深度特征聚类和语义度量的不相关标签过滤方法,其结构简单、设计合理,对庞大的样本图像数据聚类获取聚类簇,用于对样本图像数据的预分类,通过对聚类成的样本图像数据进行分析,拥有更高的有效性与正确性,同时对标签语义进行相关度度量,从而实现了不相关标签的自动过滤,可以提升深度网络的泛化性与鲁棒性。The technical problem to be solved by the present invention is to provide a method for filtering irrelevant labels based on deep feature clustering and semantic measurement, which has a simple structure and reasonable design, and can cluster huge sample image data. Obtain clusters for pre-classification of sample image data. By analyzing the clustered sample image data, it has higher validity and correctness. At the same time, it measures the relevance of label semantics, thus realizing different Automatic filtering of relevant tags can improve the generalization and robustness of deep networks.
为解决上述技术问题,本发明采用的技术方案是:基于深度特征聚类和语义度量的不相关标签过滤方法,其特征在于:包括以下步骤:In order to solve the above-mentioned technical problems, the technical solution adopted in the present invention is: an irrelevant label filtering method based on deep feature clustering and semantic measurement, characterized in that: comprising the following steps:
步骤一:传感器获取图像集X,并将图像集X存储在存储单元内,X={x1,...xi,...xn},其中xi表示第i个样本图像数据,1≤i≤n,n为正整数;Step 1: The sensor acquires the image set X, and stores the image set X in the storage unit, X={x 1 ,... xi ,...x n }, where x i represents the i-th sample image data, 1≤i≤n, n is a positive integer;
步骤二:在存储单元内建立与图像集X对应的标签集;Step 2: Establish a label set corresponding to the image set X in the storage unit;
步骤三、提取图像集图像的深度特征:对图像集X中的样本图像数据xi提取深度特征,获得深度特征φ(xi);Step 3, extract the depth features of the images in the image set: extract the depth features from the sample image data x i in the image set X, and obtain the depth features φ(xi ) ;
步骤四、对深度特征聚类获取聚类簇:以预设数目k作为类簇数对深度特征φ(xi)进行聚类,得到聚类簇集合A,A={A1,...,Af,...Ak},其中1≤f≤k,k为正整数;Step 4. Acquire clusters by clustering the deep features: cluster the deep features φ( xi ) with the preset number k as the number of clusters, and obtain the cluster set A, A={A 1 ,... ,A f ,...A k }, where 1≤f≤k, k is a positive integer;
步骤五、构建聚类簇的相关语义标签集合:根据步骤二中的原始类别标签集U获取每个聚类簇Af的聚类中心的语义标签,以聚类簇Af聚类中心的语义标签作为相关语义标签yf,得到聚类簇集合A对应的相关语义标签集合Y,Y={y1,...,yf,...yk};Step 5. Construct the relevant semantic label set of clusters: According to the original category label set U in step 2, the semantic labels of the cluster centers of each cluster A f are obtained, and the semantic labels of the cluster centers of the cluster A f are used The label is used as the relevant semantic label y f , and the relevant semantic label set Y corresponding to the cluster set A is obtained, Y={y 1 ,...,y f ,...y k };
步骤六、构建聚类簇的待度量标签集合:获取待度量标签集合P,P={P1,...,Pf,...Pk},Pf表示聚类簇Af对应的待度量标签集合,根据步骤二中的原始类别标签集U获取每个聚类簇Af下除聚类中心外的其他类别标签,将聚类簇Af下除聚类中心外的其他类别标签加入待度量标签集合Pf,t为正整数;Step 6. Construct the label set to be measured for the cluster: obtain the label set P to be measured, P={P 1 ,...,P f ,...P k }, P f represents the corresponding value of the cluster A f For the set of labels to be measured, according to the original category label set U in step 2, other category labels except the cluster center under each cluster A f are obtained, and the other category labels under the cluster A f except the cluster center Add the label set P f to be measured, t is a positive integer;
步骤七、生成语义向量:将相关语义标签集合Y和待度量标签集合P作为输入,获取相关语义标签集合Y的所有语义向量Hf,以及待度量标签集合P中Pf的所有语义向量Kfg;Step 7. Generate semantic vectors: take the relevant semantic label set Y and the label set P to be measured as input, and obtain all the semantic vectors H f of the relevant semantic label set Y and all the semantic vectors K fg of P f in the label set P to be measured ;
步骤八、计算语义向量的相关度:计算机根据公式计算相关语义标签集合Y和第f个聚类簇的待度量标签集合Pf中每个标签的相关度Simfg,其中Hf表示相关语义标签集合Y中相关语义标签yf的语义向量,Kfg表示待度量标签集合P中Pf中第g个标签的语义向量;Step 8. Calculate the relevance of the semantic vector: the computer according to the formula Calculate the correlation Sim fg of each tag in the set of related semantic tags Y and the tag set P f to be measured in the fth cluster, where H f represents the semantic vector of the related semantic tag y f in the set Y of related semantic tags, K fg represents the semantic vector of the gth label in P f in the label set P to be measured;
步骤九、根据相关度进行不相关标签过滤:将聚类簇Af中相关度Simfg低于阈值η的标签进行删除。Step 9: Filter irrelevant tags according to the correlation: delete the tags whose correlation Sim fg is lower than the threshold η in the cluster A f .
上述的一种基于深度特征聚类和语义度量的不相关标签过滤方法,其特征在于:步骤三中,利用在大型图像数据集Imagenet预训练好的深度卷积残差神经网络模型对图像集X中的样本图像数据xi提取深度特征,该网络模型由卷积层、残差层以及全连接层组成。The above-mentioned unrelated label filtering method based on deep feature clustering and semantic measurement is characterized in that: in step 3, using the deep convolutional residual neural network model pre-trained in the large image data set Imagenet to image set X The sample image data xi in extracts deep features, and the network model consists of convolutional layers, residual layers and fully connected layers.
上述的一种基于深度特征聚类和语义度量的不相关标签过滤方法,其特征在于:步骤四中,采用谱聚类算法对深度特征φ(xi)进行聚类,具体步骤包括:The aforementioned method for filtering irrelevant labels based on deep feature clustering and semantic measurement is characterized in that: in step 4, the spectral clustering algorithm is used to cluster the deep feature φ(xi ) , and the specific steps include:
步骤401:构建深度特征φ(xi)的相似度矩阵W,W为由sij组成的相似度矩阵, Step 401: Construct the similarity matrix W of the depth feature φ(xi ) , W is a similarity matrix composed of sij ,
步骤402:计算对角矩阵D,其中其中wij进行表示相似度矩阵W中第i行第j列的元素;Step 402: Calculate the diagonal matrix D, where Among them, w ij represents the element in row i and column j in the similarity matrix W;
步骤403:根据L=D-W获取深度特征φ(xi)的拉普拉斯矩阵L;Step 403: Obtain the Laplacian matrix L of the depth feature φ( xi ) according to L=DW;
步骤404:对拉普拉斯矩阵L进行特征值分解,构建特征向量空间,通过聚类算法对特征向量空间中的特征向量进行聚类,得到聚类簇集合A,A={A1,...,Af,...Ak}。Step 404: Perform eigenvalue decomposition on the Laplacian matrix L, construct an eigenvector space, and cluster the eigenvectors in the eigenvector space through a clustering algorithm to obtain a cluster set A, A={A 1 ,. ..,A f ,...A k }.
上述的一种基于深度特征聚类和语义度量的不相关标签过滤方法,其特征在于:步骤七中,生成语义向量利用近义词网络模型Synonyms。The above-mentioned irrelevant tag filtering method based on deep feature clustering and semantic measurement is characterized in that: in step 7, the semantic vector is generated using the Synonyms network model Synonyms.
上述的一种基于深度特征聚类和语义度量的不相关标签过滤方法,其特征在于:步骤九中,利用原始类别标签集U的语义向量与相关语义标签集合Y的语义向量的余弦距离,以及误标注标签集V的语义向量与相关语义标签集合Y的语义向量的余弦距离之间的差异,设定相关度阈值η。The above-mentioned unrelated label filtering method based on deep feature clustering and semantic measurement is characterized in that: in step 9, the cosine distance between the semantic vector of the original category label set U and the semantic vector of the related semantic label set Y is used, and The difference between the cosine distance between the semantic vector of the mislabeled label set V and the semantic vector of the related semantic label set Y sets the correlation threshold η.
本发明与现有技术相比具有以下优点:Compared with the prior art, the present invention has the following advantages:
1、本发明的结构简单、设计合理,实现及使用操作方便。1. The structure of the present invention is simple, the design is reasonable, and the realization, use and operation are convenient.
2、本发明采用在大型图像数据集Imagenet预训练好的深度卷积残差神经网络模型进行深度特征提取,综合了深度卷积神经网络学习能力强和残差学习收敛好的特性,特征提取与选择更具有鲁棒性,保护图像信息的完整性,提高结果性能。2. The present invention uses the deep convolutional residual neural network model pre-trained in the large-scale image data set Imagenet to carry out deep feature extraction, which combines the characteristics of strong learning ability of the deep convolutional neural network and good convergence of residual learning, feature extraction and Selection is more robust, protecting the integrity of image information and improving result performance.
3、本发明对庞大的样本图像数据聚类获取聚类簇,用于对样本图像数据的预分类,减少了人工分类所需的时间,避免了因主观差异而引起的分类结果不同,通过对聚类成的样本图像数据进行分析,可以更好地筛别图像集,拥有更高的有效性与正确性。3. The present invention clusters huge sample image data to obtain clusters for pre-classification of sample image data, reduces the time required for manual classification, and avoids different classification results caused by subjective differences. The clustered sample image data is analyzed, which can better screen the image set and have higher validity and correctness.
4、本发明分别获取相关语义标签集合Y中相关语义标签yf的语义向量和待度量标签集合Pf中第g个标签的语义向量,计算待度量标签集合Pf中第g个标签与相关语义标签集合Y中每个相关语义标签的余弦距离的平均值,作为待度量标签集合Pf中第g个标签与图像集X的相关度,以此进行相关度筛别,该方法对标签语义进行了相关度度量,从而实现了不相关标签的自动过滤。4. The present invention obtains the semantic vector of the relevant semantic label y f in the related semantic label set Y and the semantic vector of the g-th label in the set of labels to be measured P f respectively, and calculates the correlation between the g-th label in the set of labels to be measured P f The average value of the cosine distance of each related semantic label in the semantic label set Y is used as the correlation degree between the gth label in the label set P f to be measured and the image set X, so as to perform correlation screening. A correlation measure is carried out, which enables automatic filtering of irrelevant tags.
综上所述,本发明结构简单、设计合理,对庞大的样本图像数据聚类获取聚类簇,用于对样本图像数据的预分类,通过对聚类成的样本图像数据进行分析,拥有更高的有效性与正确性,同时对标签语义进行相关度度量,从而实现了不相关标签的自动过滤,可以提升深度网络的泛化性与鲁棒性。To sum up, the present invention has a simple structure and a reasonable design. It clusters huge sample image data to obtain clusters for pre-classification of sample image data. By analyzing the clustered sample image data, it has more High effectiveness and correctness, and at the same time measure the relevance of tag semantics, thereby realizing the automatic filtering of irrelevant tags, which can improve the generalization and robustness of deep networks.
下面通过附图和实施例,对本发明的技术方案做进一步的详细描述。The technical solutions of the present invention will be described in further detail below with reference to the accompanying drawings and embodiments.
附图说明Description of drawings
图1为本发明的方法流程图。Fig. 1 is a flow chart of the method of the present invention.
具体实施方式Detailed ways
下面结合附图及本发明的实施例对本发明的方法作进一步详细的说明。The method of the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments of the present invention.
需要说明的是,在不冲突的情况下,本申请中的实施例及实施例中的特征可以相互组合。下面将参考附图并结合实施例来详细说明本发明。It should be noted that, in the case of no conflict, the embodiments in the present application and the features in the embodiments can be combined with each other. The present invention will be described in detail below with reference to the accompanying drawings and examples.
需要注意的是,这里所使用的术语仅是为了描述具体实施方式,而非意图限制根据本申请的示例性实施方式。如在这里所使用的,除非上下文另外明确指出,否则单数形式也意图包括复数形式,此外,还应当理解的是,当在本说明书中使用术语“包含”和/或“包括”时,其指明存在特征、步骤、操作、器件、组件和/或它们的组合。It should be noted that the terminology used here is only for describing specific implementations, and is not intended to limit the exemplary implementations according to the present application. As used herein, unless the context clearly dictates otherwise, the singular is intended to include the plural, and it should also be understood that when the terms "comprising" and/or "comprising" are used in this specification, they mean There are features, steps, operations, means, components and/or combinations thereof.
需要说明的是,本申请的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便这里描述的本申请的实施方式例如能够以除了在这里图示或描述的那些以外的顺序实施。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元的过程、方法、系统、产品或设备不必限于清楚地列出的那些步骤或单元,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它步骤或单元。It should be noted that the terms "first" and "second" in the description and claims of the present application and the above drawings are used to distinguish similar objects, but not necessarily used to describe a specific sequence or sequence. It is to be understood that the data so used are interchangeable under appropriate circumstances such that the embodiments of the application described herein, for example, can be practiced in sequences other than those illustrated or described herein. Furthermore, the terms "comprising" and "having", as well as any variations thereof, are intended to cover a non-exclusive inclusion, for example, a process, method, system, product or device comprising a sequence of steps or elements is not necessarily limited to the expressly listed instead, may include other steps or elements not explicitly listed or inherent to the process, method, product or apparatus.
为了便于描述,在这里可以使用空间相对术语,如“在……之上”、“在……上方”、“在……上表面”、“上面的”等,用来描述如在图中所示的一个器件或特征与其他器件或特征的空间位置关系。应当理解的是,空间相对术语旨在包含除了器件在图中所描述的方位之外的在使用或操作中的不同方位。例如,如果附图中的器件被倒置,则描述为“在其他器件或构造上方”或“在其他器件或构造之上”的器件之后将被定位为“在其他器件或构造下方”或“在其他器件或构造之下”。因而,示例性术语“在……上方”可以包括“在……上方”和“在……下方”两种方位。该器件也可以其他不同方式定位(旋转90度或处于其他方位),并且对这里所使用的空间相对描述作出相应解释。For the convenience of description, spatially relative terms may be used here, such as "on ...", "over ...", "on the surface of ...", "above", etc., to describe the The spatial positional relationship between one device or feature shown and other devices or features. It will be understood that the spatially relative terms are intended to encompass different orientations of the device in use or operation in addition to the orientation depicted in the figures. For example, if the device in the figures is turned over, devices described as "above" or "above" other devices or configurations would then be oriented "beneath" or "above" the other devices or configurations. under other devices or configurations”. Thus, the exemplary term "above" can encompass both an orientation of "above" and "beneath". The device may be otherwise oriented (rotated 90 degrees or at other orientations) and the spatially relative descriptions used herein interpreted accordingly.
如图1所示,本发明包括以下步骤:As shown in Figure 1, the present invention comprises the following steps:
步骤一:传感器获取图像集X,并将图像集X存储在存储单元内,X={x1,...xi,...xn},其中xi表示第i个样本图像数据,1≤i≤n,n为正整数。Step 1: The sensor acquires the image set X, and stores the image set X in the storage unit, X={x 1 ,... xi ,...x n }, where x i represents the i-th sample image data, 1≤i≤n, n is a positive integer.
实际使用时,通过传感器采集不同类型的样本图像数据,对象不同,传感器所采集的样本图像数据也有所不同。In actual use, different types of sample image data are collected by the sensor, and the sample image data collected by the sensor is also different for different objects.
步骤二:在存储单元内建立与图像集X对应的标签集。具体实施时,数据集包括图像集X和标签集,标签集包括原始类别标签集U和误标注标签集V,U={u1,...,up,...uh},V={v1,...,vq,...vl},其中,1≤p≤h,1≤q≤l,h、l均为正整数,l+h=n。图像集X中每一个样本图像数据xi都对应原始类别标签集U中的一个标签。误标注标签集V用于存储最后筛选出来的不相关标签。Step 2: Establish a label set corresponding to the image set X in the storage unit. During specific implementation, the data set includes an image set X and a label set, and the label set includes the original category label set U and the mislabeled label set V, U={u 1 ,...,up p ,...u h }, V ={v 1 ,...,v q ,...v l }, wherein, 1≤p≤h, 1≤q≤l, h and l are both positive integers, and l+h=n. Each sample image data xi in the image set X corresponds to a label in the original category label set U. The mislabeled label set V is used to store the last screened out irrelevant labels.
步骤三、获取图像深度特征:计算机对图像集X中的样本图像数据xi提取深度特征,获得深度特征φ(xi)。Step 3: Obtain image depth features: the computer extracts depth features from the sample image data x i in image set X to obtain depth features φ(xi ) .
具体实施时,为了提升聚类效果,需要将图像集X中的样本图像数据xi转化为一个适当的特征表示。具体实施时,计算机采用大型图像数据集Imagenet预训练深度卷积残差神经网络,根据所采集的传感器数据提取出有关样本图像数据xi的深度特征φ(xi),综合了深度卷积神经网络学习能力强和残差学习收敛好的特性,特征提取与选择更具有鲁棒性,解决了图像特征细节缺少的问题,保护图像信息的完整性,提高了结果性能。During specific implementation, in order to improve the clustering effect, it is necessary to convert the sample image data xi in the image set X into an appropriate feature representation. In the specific implementation, the computer uses the large-scale image data set Imagenet to pre-train the deep convolutional residual neural network, and extracts the depth feature φ( xi ) of the sample image data x i according to the collected sensor data, and integrates the deep convolutional neural network With strong network learning ability and good convergence of residual learning, feature extraction and selection are more robust, which solves the problem of lack of image feature details, protects the integrity of image information, and improves the performance of results.
具体为:深度卷积神经网络模型利用每层的卷积核对样本图像数据xi做卷积操作,并提取每个样本图像数据xi的初始特征,使用残差网络模型在深度卷积神经网络添加跳跃连接,可将样本图像数据xi的初始特征直接传输到深度卷积神经网络中后层,提高结果性能,然后将特征输入模型的全连接层进行拼接,得到样本图像数据xi的深度特征φ(xi),深度特征φ(xi)包含着和样本图像数据有关xi的特征,因此深度特征φ(xi)影响最终的分类过滤效果。Specifically: the deep convolutional neural network model uses the convolution kernel of each layer to perform convolution operations on the sample image data xi , and extracts the initial features of each sample image data xi , and uses the residual network model in the deep convolutional neural network Adding a skip connection can directly transfer the initial features of the sample image data xi to the back layer of the deep convolutional neural network to improve the performance of the result, and then stitch the features into the fully connected layer of the model to obtain the depth of the sample image data xi Feature φ( xi ), depth feature φ(xi ) contains features related to sample image data xi , so depth feature φ( xi ) affects the final classification and filtering effect.
步骤四、获取聚类簇:以预设数目k作为类簇数对深度特征φ(xi)进行聚类,得到聚类簇集合A,A={A1,...,Af,...Ak},其中1≤f≤k,k为正整数;Step 4. Obtain clusters: cluster the depth features φ( xi ) with the preset number k as the number of clusters, and obtain a cluster set A, A={A 1 ,...,A f ,. ..A k }, where 1≤f≤k, k is a positive integer;
谱聚类算法以相似度矩阵为基础,将普通聚类问题转化为图的划分问题,谱聚类算法建立在谱图理论基础上,在聚类时不受样本空间形状的限制,因此优于传统的聚类算法。谱聚类算法在求解时从全局出发,具有收敛于全局最优解的优点,不会陷入局部最优解,同时可以保证不同类间的相似度最小,同一类内的相似度最大,其性能及适用场景优于传统的聚类算法。因此本申请优选采用谱聚类算法对深度特征φ(xi)进行聚类。The spectral clustering algorithm is based on the similarity matrix, and transforms the ordinary clustering problem into a graph division problem. The spectral clustering algorithm is based on the spectral graph theory, and it is not limited by the shape of the sample space when clustering, so it is better than Traditional clustering algorithms. The spectral clustering algorithm starts from the whole world when solving, and has the advantage of converging on the global optimal solution without falling into the local optimal solution. At the same time, it can ensure the minimum similarity between different classes and the largest similarity within the same class. Its performance And applicable scenarios are better than traditional clustering algorithms. Therefore, the present application preferably uses a spectral clustering algorithm to cluster the depth features φ(xi ) .
采用谱聚类算法进行聚类的具体过程为:构建图像集X的相似度矩阵W,W为由sij组成的相似度矩阵,σ表示标准差。相似度矩阵W表示为W=(wij)i,j=1,...n,计算机根据公式计算对角矩阵D,D={d1,...di,...dn}。根据L=D-W获取深度特征φ(xi)的拉普拉斯矩阵L,对拉普拉斯矩阵L进行特征值分解,构建特征向量空间,通过聚类算法对特征向量空间中的特征向量进行聚类,得到聚类簇集合A,A={A1,...,Af,...Ak}。The specific process of clustering using the spectral clustering algorithm is: constructing the similarity matrix W of the image set X, W is a similarity matrix composed of sij , σ represents the standard deviation. The similarity matrix W is expressed as W=(w ij ) i, j=1,...n , and the computer according to the formula Calculate the diagonal matrix D, D={d 1 ,...d i ,...d n }. Obtain the Laplacian matrix L of the depth feature φ( xi ) according to L=DW, perform eigenvalue decomposition on the Laplacian matrix L, construct the eigenvector space, and perform clustering algorithm on the eigenvectors in the eigenvector space Clustering to obtain a cluster set A, A={A 1 ,...,A f ,...A k }.
本申请对庞大的样本图像数据聚类获取聚类簇,用于对样本图像数据的预分类,减少了人工分类所需的时间,避免了因主观差异而引起的分类结果不同,通过对聚类成的样本图像数据进行分析,可以更好地筛别图像集,拥有更高的有效性与正确性,为不相关标签的过滤提供更可靠地方法。This application clusters huge sample image data to obtain cluster clusters for pre-classification of sample image data, reduces the time required for manual classification, and avoids different classification results caused by subjective differences. Analyzing the generated sample image data can better screen the image set, have higher validity and correctness, and provide a more reliable method for filtering irrelevant labels.
步骤五、构建相关语义标签集合:根据步骤二中的原始类别标签集U获取每个聚类簇Af的聚类中心的语义标签,以聚类簇Af聚类中心的语义标签作为相关语义标签yf,得到相关语义标签集合Y,Y={y1,...,yf,...yk}。Step 5. Construct the relevant semantic label set: According to the original category label set U in step 2, obtain the semantic label of the cluster center of each cluster A f , and use the semantic label of the cluster center of the cluster A f as the relevant semantic label y f , to obtain a set of related semantic labels Y, Y={y 1 ,...,y f ,...y k }.
由于图像集X中每一个样本图像数据xi都对应原始类别标签集U中的一个标签,因此聚类簇Af的聚类中心也对应原始类别标签集U中的一个语义标签。Since each sample image data xi in the image set X corresponds to a label in the original category label set U, the cluster center of the cluster A f also corresponds to a semantic label in the original category label set U.
步骤六、构建待度量标签集合:获取待度量标签集合P,P={P1,...,Pf,...Pk},Pf表示聚类簇Af对应的待度量标签集合,将聚类簇Af下除聚类中心外的聚类元素所对应的标签组合起来,即构成待度量标签集合Pf,t为正整数。Step 6. Build a set of labels to be measured: obtain a set of labels to be measured P, P={P 1 ,...,P f ,...P k }, P f represents the set of labels to be measured corresponding to the cluster A f , combine the labels corresponding to the cluster elements under the cluster A f except the cluster center to form the label set P f to be measured, t is a positive integer.
同理,由于图像集X中每一个样本图像数据xi都对应原始类别标签集U中的一个标签,因此聚类簇Af下除聚类中心外的聚类元素也各自对应原始类别标签集U中的一个语义标签。Similarly, since each sample image data xi in the image set X corresponds to a label in the original category label set U, the cluster elements under the cluster A f except the cluster center also correspond to the original category label set A semantic label in U.
步骤七、生成语义向量:将相关语义标签集合Y和待度量标签集合P作为输入,获取相关语义标签集合Y的所有语义向量Hf,以及待度量标签集合P中Pf的所有语义向量Kfg。Step 7. Generate semantic vectors: take the relevant semantic label set Y and the label set P to be measured as input, and obtain all the semantic vectors H f of the relevant semantic label set Y and all the semantic vectors K fg of P f in the label set P to be measured .
需要说明的是,本申请优选采用近义词网络模型Synonyms。近义词网络模型Synonyms是一种训练好的word2vec模型,word2vec使用大量数据,利用上下文信息进行训练,将词汇映射到低维空间,在算法层面上,检索更是基于了“距离”而非“匹配”,基于“语义”而非“形式”。近义词网络模型Synonyms作为训练好的word2vec模型,可将每个词映射到一个向量,可用来表示词对词之间的关系,具备度量词语和词语之间相关度的能力。It should be noted that the present application preferably adopts the Synonyms network model. The Synonyms network model is a well-trained word2vec model. word2vec uses a large amount of data, uses contextual information for training, and maps vocabulary to low-dimensional space. At the algorithm level, retrieval is based on "distance" rather than "match". , based on "semantics" rather than "form". The synonym network model Synonyms, as a trained word2vec model, can map each word to a vector, which can be used to represent the relationship between words and words, and has the ability to measure the correlation between words and words.
向近义词网络模型Synonyms输入词语进行预测,近义词网络模型Synonyms输出隐层变量,根据隐层变量计算出的参数就是这个词语对应的的语义向量。换言之,近义词网络模型Synonyms可以根据输入的词语,输出这个词语的数学表达形式,就是语义向量。Input words to the Synonyms network model for prediction, and the Synonyms network model Synonyms outputs hidden layer variables, and the parameters calculated according to the hidden layer variables are the semantic vectors corresponding to the words. In other words, the synonym network model Synonyms can output the mathematical expression of the word according to the input word, which is the semantic vector.
步骤八、计算语义向量的相关度:计算机根据公式计算相关语义标签集合Y和第f个聚类簇的待度量标签集合Pf中每个标签的相关度Simfg,其中Hf表示相关语义标签集合Y中相关语义标签yf的语义向量,Kfg表示待度量标签集合P中Pf中第g个标签的语义向量;Step 8. Calculate the relevance of the semantic vector: the computer according to the formula Calculate the correlation Sim fg of each tag in the set of related semantic tags Y and the tag set P f to be measured in the fth cluster, where H f represents the semantic vector of the related semantic tag y f in the set Y of related semantic tags, K fg represents the semantic vector of the gth label in P f in the label set P to be measured;
Simfg表示待度量标签集合Pf中第g个标签与相关语义标签集合Y中每个相关语义标签的余弦距离的平均值,因此将Simfg作为待度量标签集合Pf中第g个标签与图像集X的相关度,以此作为待度量标签集合Pf中第g个标签是否应该被过滤的指标。Sim fg represents the average value of the cosine distance between the gth tag in the tag set P f to be measured and each related semantic tag in the related semantic tag set Y, so Sim fg is used as the gth tag in the tag set P f to be measured and The correlation degree of the image set X is used as an indicator of whether the gth label in the label set P f to be measured should be filtered.
步骤九、根据相关度进行不相关标签过滤:在原始类别标签集U中,将聚类簇Af中相关度Simfg低于阈值η的标签进行删除;在图像集X中,将聚类簇Af中相关度Simfg低于阈值η的标签所对应的样本图像数据xi进行删除,从而得到可训练的数据集,降低了深度学习模型参数的拟合时间,提高了拟合效率,使用效果好。Step 9. Filter irrelevant labels according to the correlation: In the original category label set U, delete the labels whose correlation Sim fg is lower than the threshold η in the cluster A f ; in the image set X, delete the cluster cluster A f The sample image data xi corresponding to the label whose correlation Sim fg is lower than the threshold η in A f is deleted, thereby obtaining a trainable data set, reducing the fitting time of the deep learning model parameters, and improving the fitting efficiency. The effect is good.
具体实施时,利用原始类别标签集U的语义向量与相关语义标签集合Y的语义向量的余弦距离,以及误标注标签集V的语义向量与相关语义标签集合Y的语义向量的余弦距离的差异,设定相关度阈值η。During specific implementation, the cosine distance between the semantic vector of the original category label set U and the semantic vector of the related semantic label set Y, and the cosine distance between the semantic vector of the mislabeled label set V and the semantic vector of the related semantic label set Y are used, Set the correlation threshold η.
以上所述,仅是本发明的实施例,并非对本发明作任何限制,凡是根据本发明技术实质对以上实施例所作的任何简单修改、变更以及等效结构变化,均仍属于本发明技术方案的保护范围内。The above is only an embodiment of the present invention, and does not limit the present invention in any way. Any simple modifications, changes and equivalent structural changes made to the above embodiments according to the technical essence of the present invention still belong to the technical solution of the present invention. within the scope of protection.
Claims (5)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010992837.4A CN112232374B (en) | 2020-09-21 | 2020-09-21 | Irrelevant label filtering method based on depth feature clustering and semantic measurement |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010992837.4A CN112232374B (en) | 2020-09-21 | 2020-09-21 | Irrelevant label filtering method based on depth feature clustering and semantic measurement |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112232374A CN112232374A (en) | 2021-01-15 |
CN112232374B true CN112232374B (en) | 2023-04-07 |
Family
ID=74108089
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010992837.4A Active CN112232374B (en) | 2020-09-21 | 2020-09-21 | Irrelevant label filtering method based on depth feature clustering and semantic measurement |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112232374B (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113111180B (en) * | 2021-03-22 | 2022-01-25 | 杭州祺鲸科技有限公司 | Chinese medical synonym clustering method based on deep pre-training neural network |
CN113435308B (en) * | 2021-06-24 | 2023-05-30 | 平安国际智慧城市科技股份有限公司 | Text multi-label classification method, device, equipment and storage medium |
CN114494654B (en) * | 2021-12-30 | 2025-02-14 | 杭州群核信息技术有限公司 | Label positioning method, device, computer equipment and storage medium for question semantic graph |
CN114528844B (en) * | 2022-01-14 | 2024-09-06 | 中国平安人寿保险股份有限公司 | Intention recognition method, device, computer equipment and storage medium |
CN114596434B (en) * | 2022-03-15 | 2024-10-15 | 山东财经大学 | Image block and label matching method and system based on three-part graph model |
CN114998634B (en) * | 2022-08-03 | 2022-11-15 | 广州此声网络科技有限公司 | Image processing method, image processing device, computer equipment and storage medium |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017084267A1 (en) * | 2015-11-18 | 2017-05-26 | 乐视控股(北京)有限公司 | Method and device for keyphrase extraction |
CN111080551A (en) * | 2019-12-13 | 2020-04-28 | 太原科技大学 | Multi-label image completion method based on deep convolutional features and semantic neighbors |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120158686A1 (en) * | 2010-12-17 | 2012-06-21 | Microsoft Corporation | Image Tag Refinement |
CN103092911B (en) * | 2012-11-20 | 2016-02-03 | 北京航空航天大学 | A kind of mosaic society label similarity is based on the Collaborative Filtering Recommendation System of k nearest neighbor |
US9082047B2 (en) * | 2013-08-20 | 2015-07-14 | Xerox Corporation | Learning beautiful and ugly visual attributes |
US20150347562A1 (en) * | 2014-06-02 | 2015-12-03 | Qualcomm Incorporated | Deriving user characteristics from users' log files |
US20180300315A1 (en) * | 2017-04-14 | 2018-10-18 | Novabase Business Solutions, S.A. | Systems and methods for document processing using machine learning |
US10482323B2 (en) * | 2017-08-22 | 2019-11-19 | Autonom8, Inc. | System and method for semantic textual information recognition |
CN107563444A (en) * | 2017-09-05 | 2018-01-09 | 浙江大学 | A kind of zero sample image sorting technique and system |
RU2711125C2 (en) * | 2017-12-07 | 2020-01-15 | Общество С Ограниченной Ответственностью "Яндекс" | System and method of forming training set for machine learning algorithm |
US11194842B2 (en) * | 2018-01-18 | 2021-12-07 | Samsung Electronics Company, Ltd. | Methods and systems for interacting with mobile device |
CN111177444A (en) * | 2020-01-02 | 2020-05-19 | 杭州创匠信息科技有限公司 | Image marking method and electronic equipment |
-
2020
- 2020-09-21 CN CN202010992837.4A patent/CN112232374B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017084267A1 (en) * | 2015-11-18 | 2017-05-26 | 乐视控股(北京)有限公司 | Method and device for keyphrase extraction |
CN111080551A (en) * | 2019-12-13 | 2020-04-28 | 太原科技大学 | Multi-label image completion method based on deep convolutional features and semantic neighbors |
Non-Patent Citations (1)
Title |
---|
李艳 ; 贾君枝 ; .基于向量空间模型的标签树构建方法研究.情报学报.2014,(03),全文. * |
Also Published As
Publication number | Publication date |
---|---|
CN112232374A (en) | 2021-01-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112232374B (en) | Irrelevant label filtering method based on depth feature clustering and semantic measurement | |
CN110414368B (en) | Unsupervised pedestrian re-identification method based on knowledge distillation | |
CN113065409B (en) | An unsupervised person re-identification method based on camera distribution difference alignment constraint | |
Jiao et al. | SAR images retrieval based on semantic classification and region-based similarity measure for earth observation | |
Zhang et al. | Detecting densely distributed graph patterns for fine-grained image categorization | |
CN114358188B (en) | Feature extraction model processing, sample retrieval method, device and computer equipment | |
CN108108657A (en) | A kind of amendment local sensitivity Hash vehicle retrieval method based on multitask deep learning | |
Bu et al. | 3D shape recognition and retrieval based on multi-modality deep learning | |
CN112307995A (en) | A semi-supervised person re-identification method based on feature decoupling learning | |
EP3166020A1 (en) | Method and apparatus for image classification based on dictionary learning | |
CN101587478A (en) | Methods and devices for training, automatically labeling and searching images | |
CN114676777B (en) | Self-supervision learning fine-granularity image classification method based on twin network | |
CN112199532A (en) | Zero sample image retrieval method and device based on Hash coding and graph attention machine mechanism | |
CN114926742B (en) | A loop detection and optimization method based on second-order attention mechanism | |
CN107977661A (en) | The region of interest area detecting method decomposed based on full convolutional neural networks and low-rank sparse | |
Al-Jubouri | Content-based image retrieval: Survey | |
CN105205135A (en) | 3D (three-dimensional) model retrieving method based on topic model and retrieving device thereof | |
CN111882000A (en) | A network structure and method applied to small sample fine-grained learning | |
CN115063831A (en) | A high-performance pedestrian retrieval and re-identification method and device | |
CN110705384B (en) | Vehicle re-identification method based on cross-domain migration enhanced representation | |
CN119202934A (en) | A multimodal annotation method based on deep learning | |
Yu et al. | A light-weighted hypergraph neural network for multimodal remote sensing image retrieval | |
Oussama et al. | A fast weighted multi-view Bayesian learning scheme with deep learning for text-based image retrieval from unlabeled galleries | |
CN115601648A (en) | Open set image classification method based on cluster analysis | |
Manisha et al. | Content-based image retrieval through semantic image segmentation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
OL01 | Intention to license declared | ||
OL01 | Intention to license declared |