[go: up one dir, main page]

CN104657744A - Multi-classifier training method and classifying method based on non-deterministic active learning - Google Patents

Multi-classifier training method and classifying method based on non-deterministic active learning Download PDF

Info

Publication number
CN104657744A
CN104657744A CN201510046879.8A CN201510046879A CN104657744A CN 104657744 A CN104657744 A CN 104657744A CN 201510046879 A CN201510046879 A CN 201510046879A CN 104657744 A CN104657744 A CN 104657744A
Authority
CN
China
Prior art keywords
sample
classifier
unlabeled
information
sample set
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510046879.8A
Other languages
Chinese (zh)
Other versions
CN104657744B (en
Inventor
张晓宇
王树鹏
吴广君
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Information Engineering of CAS
Original Assignee
Institute of Information Engineering of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Information Engineering of CAS filed Critical Institute of Information Engineering of CAS
Priority to CN201510046879.8A priority Critical patent/CN104657744B/en
Publication of CN104657744A publication Critical patent/CN104657744A/en
Application granted granted Critical
Publication of CN104657744B publication Critical patent/CN104657744B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

本发明公开了一种基于非确定主动学习的多分类器训练方法及分类方法。本方法为:1)选取或初始化一多分类器;对未标注样本集中的每一样本,利用该多分类器计算该样本的总体信息量Info;所述总体信息量为:模型变更信息量与模型调优信息量之和;2)对该未标注样本集进行聚类,得到J个子类;3)从每个子类中选取总体信息量Info值最小的若干未标注样本;再从所选样本中选取K个样本进行标注后加入到已标注样本集L;4)将更新后的已标注集L作为训练数据重新训练该多分类器;5)迭代执行步骤1)~4)设定次数;然后利用最终得到的多分类器对未标注集进行分类。本发明实现样本信息量的综合评价,从而获得高效化、智能化的多分类器。

The invention discloses a multi-classifier training method and classification method based on non-deterministic active learning. The method is: 1) select or initialize a multi-classifier; for each sample in the unmarked sample set, use the multi-classifier to calculate the overall information amount Info of the sample; the overall information amount is: model change information amount and The sum of model tuning information; 2) cluster the unlabeled sample set to obtain J subclasses; 3) select some unlabeled samples with the smallest overall information Info value from each subclass; Select K samples from among them for labeling and add them to the labeled sample set L; 4) Use the updated labeled set L as training data to retrain the multi-classifier; 5) Iteratively execute steps 1) to 4) to set the number of times; The resulting multi-classifier is then used to classify the unlabeled set. The invention realizes the comprehensive evaluation of the amount of sample information, thereby obtaining a highly efficient and intelligent multi-classifier.

Description

一种基于非确定主动学习的多分类器训练方法及分类方法A multi-classifier training method and classification method based on non-deterministic active learning

技术领域technical field

本发明涉及一种基于非确定主动学习的多分类器训练方法及分类方法,属于软件工程技术领域。The invention relates to a multi-classifier training method and classification method based on non-deterministic active learning, belonging to the technical field of software engineering.

背景技术Background technique

数据分类一直是人们的研究热点,比如专利ZL 201010166225.6“一种基于在线学习的自适应级联分类器训练方法”,专利ZL 200910076428.3“一种跨领域的文本情感分类器的训练方法和分类方法”,专利ZL 200810094208.9“文档分类器生成方法和系统”。Data classification has always been a research hotspot, such as patent ZL 201010166225.6 "an adaptive cascade classifier training method based on online learning", patent ZL 200910076428.3 "a cross-domain text sentiment classifier training method and classification method" , Patent ZL 200810094208.9 "Document Classifier Generation Method and System".

在海量数据的分类问题中,“主动学习”(参考文献:McCallum and K.Nigam,“EmployingEM in pool-based active learning for text classification,”in Proc.of the 15th InternationalConference on Machine Learning,1998,pp.350–358.)是一种高效利用专家标注的机器学习方法,其主要思想是:由机器主动地、有针对性地选择最有信息的样本交给专家进行标注(向专家提出查询),从而在有限的样本标注量下获得尽可能大的分类性能提升,比如参考授权专利:ZL 201210050383“基于主动学习和半监督学习的多类图像分类方法”;ZL200810082814.9“用于使提升分类器适合于新样本的方法”。In the classification problem of massive data, "active learning" (reference: McCallum and K. Nigam, "EmployingEM in pool-based active learning for text classification," in Proc. of the 15th International Conference on Machine Learning, 1998, pp. 350–358.) is a machine learning method that efficiently utilizes expert labeling. Its main idea is: the machine actively and targetedly selects the most informative samples and sends them to experts for labeling (query to experts), so that Obtain the largest classification performance improvement under the limited amount of sample labels, such as referring to authorized patents: ZL 201210050383 "Multi-class image classification method based on active learning and semi-supervised learning"; ZL200810082814.9 "Used to make the boost classifier suitable for method for new samples".

在样本标注代价高且数量有限、而未标注样本多且易于获得的应用情境下,主动学习的优势尤为明显。选择性采样策略是主动学习的关键环节。现有选择性采样策略大致包括以下几种类型——(1)基于不确定度:将其当前模型最不确定如何进行分类的样本提交给专家标注(参考文献:D.Lewis and W.Gale,“A sequential algorithm for training text classifiers,”In Proc.of the ACM SIGIR Conference on Research and Development in Information Retrieval,1994,pp.3–12.);(2)基于组合决策:从不同模型出发,采用投票模式,将分歧最大的样本提交给专家标注(参考文献:H.S.Seung,M.Opper,and H.Sompolinsky,“Query by committee,”In Proc.oftheACM Workshop on Computational Learning Theory,1992,pp.287–294);(3)基于期望误差最小化:从决策理论出发,估计未标注样本被标注之后模型的期望误差,最终选择能够获得最小期望误差的样本提交给专家标注(参考文献:Y.Guo and R.Greiner,“Optimistic activelearning using mutual information,”In Proc.of International Joint Conference on ArtificialIntelligence,2007,pp.823–829.)。The advantages of active learning are particularly obvious in the application scenarios where the cost of sample labeling is high and the number is limited, while the number of unlabeled samples is large and easy to obtain. Selective sampling strategies are a key part of active learning. Existing selective sampling strategies roughly include the following types—(1) Uncertainty-based: Submit the samples whose current model is most uncertain about how to classify them to experts for labeling (References: D.Lewis and W.Gale, "A sequential algorithm for training text classifiers," In Proc.of the ACM SIGIR Conference on Research and Development in Information Retrieval, 1994, pp.3–12.); (2) Combination-based decision-making: starting from different models, using voting mode, and submit the most divergent samples to expert annotation (references: H.S.Seung, M.Opper, and H.Sompolinsky, “Query by committee,” In Proc. of the ACM Workshop on Computational Learning Theory, 1992, pp.287–294 ); (3) Based on the minimization of expected error: starting from the decision theory, estimate the expected error of the model after the unlabeled sample is marked, and finally select the sample that can obtain the minimum expected error and submit it to the expert for labeling (references: Y.Guo and R . Greiner, “Optimistic active learning using mutual information,” In Proc. of International Joint Conference on Artificial Intelligence, 2007, pp.823–829.).

本发明文档的符号表示如下:样本用特征向量x表示;标注用y∈CN={1,2,...,N}表示,其中N表示类别数目;未标注集和已标注集分别用U和L表示;分类模型用后验概率表示,其中表示对应于已标注集L的N分类模型的参数。The notation of the document of the present invention is as follows: the sample is represented by the feature vector x ; Represented by U and L; the classification model uses the posterior probability said, among them Denotes the parameters of the N classification model corresponding to the labeled set L.

传统主动学习方法中,类别数目N可以通过经验分析或先验知识预先获知,从而视为常量,这类方法称为“确定主动学习”(DeterminateActive Learning,简称D-AL)。根据类别数目的不同(N=2或N>2),相应的分类模型可以划分为两种——二分类(binary)和多分类(multi-class)。二分类模型将样本分到两类中的一类,是一种被广泛研究和应用的基本分类模型;多分类模型将样本分到多类中的一类,是二分类模型的一般化形式。多分类模型的构建方式包括以下两种:In traditional active learning methods, the number of categories N can be known in advance through empirical analysis or prior knowledge, so it can be regarded as a constant. This type of method is called "Determinate Active Learning" (D-AL for short). According to the number of categories (N=2 or N>2), the corresponding classification models can be divided into two types—binary and multi-class. The binary classification model divides samples into one of two categories, which is a basic classification model that has been widely studied and applied; the multi-classification model divides samples into one of multiple categories, which is a generalized form of the binary classification model. There are two ways to build a multi-classification model:

(1)一种直接的处理方法是将多分类模型转化为多个二分类模型。在训练阶段,针对每个类别或每两个类别对,构建相应的二分类模型。在预测阶段,训练出的多个二分类模型通过投票或融合的方式组合成一个总的分类模型。例如,对于每个类别c∈CN,利用标注yc={0,1}指示样本x是否属于该类,逻辑回归可以用于二分类模型的构建:(1) A straightforward approach is to transform the multi-classification model into multiple binary classification models. In the training phase, for each category or every pair of two categories, a corresponding binary classification model is constructed. In the prediction stage, multiple trained binary classification models are combined into a total classification model by voting or fusion. For example, for each category c∈C N , using the label y c ={0,1} to indicate whether the sample x belongs to this category, logistic regression can be used for the construction of a binary classification model:

P ( y c = 1 | x ; θ L c 2 ) = θ θ L c 2 ( x ) = 1 1 + exp ( - ( θ L c 2 ) T x ) .     公式(1) P ( the y c = 1 | x ; θ L c 2 ) = θ θ L c 2 ( x ) = 1 1 + exp ( - ( θ L c 2 ) T x ) . Formula 1)

最终的预测标注为后验概率最大的类别:The final prediction is labeled as the category with the largest posterior probability:

y * = arg max c ∈ C N P ( y c = 1 | x ; θ L c 2 ) .     公式(2) the y * = arg max c ∈ C N P ( the y c = 1 | x ; θ L c 2 ) . Formula (2)

(2)另一种处理方法将各类别综合考虑、统一建模。例如,给定样本x,softmax回归用一个N维向量估计标注y取1~N中每一个值的概率,从而在一个统一的过程中同时对N个类别进行建模:(2) Another processing method considers all categories comprehensively and models them in a unified manner. For example, given a sample x, softmax regression uses an N-dimensional vector to estimate the probability of labeling y taking each value from 1 to N, thereby simultaneously modeling N categories in a unified process:

P ( y = 1 | x ; θ L N ) P ( y = 2 | x ; θ L N ) . . . P ( y = N | x ; θ L N ) = h θ L N ( x ) = 1 Σ i = 1 N exp ( ( θ L N ) i T x ) exp ( ( θ L N ) 1 T x ) exp ( ( θ L N ) 2 T x ) . . . exp ( ( θ L N ) N T x ) .     公式(3) P ( the y = 1 | x ; θ L N ) P ( the y = 2 | x ; θ L N ) . . . P ( the y = N | x ; θ L N ) = h θ L N ( x ) = 1 Σ i = 1 N exp ( ( θ L N ) i T x ) exp ( ( θ L N ) 1 T x ) exp ( ( θ L N ) 2 T x ) . . . exp ( ( θ L N ) N T x ) . Formula (3)

由于样本属于各类的概率分布进行统一建模,结果具有直接可比性,向量中最大的元素对应于最终的预测标注,因此该处理方法更加适用于多分类模型构建。Since the probability distribution of samples belonging to various types is modeled uniformly, the results are directly comparable, and the largest element in the vector corresponds to the final prediction label, so this processing method is more suitable for multi-classification model construction.

为了优化分类模型,传统的基于确定主动学习的多分类方法选取最有信息的样本提交给专家进行标注,从而实现模型更新。最有信息样本的选取方法为:分别计算未标注集中每个样本在标注后模型的期望误差,选取最小化期望误差的样本作为最有信息样本,公式化表示如下:In order to optimize the classification model, the traditional multi-classification method based on deterministic active learning selects the most informative samples and submits them to experts for labeling, so as to realize the model update. The selection method of the most informative sample is: respectively calculate the expected error of each sample in the unlabeled set after labeling, and select the sample that minimizes the expected error as the most informative sample. The formula is as follows:

x D - AL * = arg min x ∈ U Σ y ~ ∈ C N P ( y ~ | x ; θ L N ) F ( x , y ~ ; θ L N ) .     公式(4) x D. - AL * = arg min x ∈ u Σ the y ~ ∈ C N P ( the y ~ | x ; θ L N ) f ( x , the y ~ ; θ L N ) . Formula (4)

其中,in,

F ( x , y ~ ; θ L N ) = Σ x u ∈ U - x H ( y u | x u ; θ L + ( x , y ~ ) N ) = Σ x u ∈ U - x ( - Σ y ~ u ∈ C N P ( y ~ u | x u ; θ L + ( x , y ~ ) N ) · log P ( y ~ u | x u ; θ L + ( x , y ~ ) N ) ) .     公式(5) f ( x , the y ~ ; θ L N ) = Σ x u ∈ u - x h ( the y u | x u ; θ L + ( x , the y ~ ) N ) = Σ x u ∈ u - x ( - Σ the y ~ u ∈ C N P ( the y ~ u | x u ; θ L + ( x , the y ~ ) N ) &Center Dot; log P ( the y ~ u | x u ; θ L + ( x , the y ~ ) N ) ) . Formula (5)

公式表示在给定现有模型参数和新标注样本的情况下,其它未标注样本xu∈U-x的信息熵之和;表示样本被标注之后新的已标注集。The formula expresses that given the existing model parameters and the newly annotated sample In the case of , the information entropy of other unlabeled samples x u ∈ Ux Sum; Indicates the sample The new labeled set after being labeled.

根据公式选取出的最有信息样本通过人机交互的形式由专家进行标注,标注完成后该样本从未标注集去除并加入已标注集。The most informative sample selected according to the formula Annotated by experts in the form of human-computer interaction, after the annotation is completed, the sample is removed from the unlabeled set and added to the labeled set.

对于基于确定主动学习的多分类方法,类别数目N事先已知,因而可以据此直接确定N分类模型的形式,剩下的任务是通过确定主动学习选取和标注最有信息样本,从而在现有模型框架下不断优化模型参数然而,在很多实际应用中,类别数目往往无法事先准确获知;甚至在有些应用中,类别数目会随着时间推移不断变化。在上述情况下,类别数目N本身为需要求解的变量,主动学习不仅需要优化现有模型参数同时还需要根据样本分布更新类别数目N(即分类模型的形式)。For the multi-classification method based on deterministic active learning, the number of categories N is known in advance, so the form of the N classification model can be directly determined accordingly. The remaining task is to select and label the most informative samples through deterministic active learning, so that in the existing Continuously optimize model parameters under the model framework However, in many practical applications, the number of categories cannot be accurately known in advance; even in some applications, the number of categories will change over time. In the above cases, the number of categories N itself is the variable that needs to be solved, active learning not only needs to optimize the existing model parameters At the same time, the number of categories N (that is, the form of the classification model) needs to be updated according to the sample distribution.

为表述清楚起见,本发明文档中将现有N分类模型下模型参数的优化称为“模型调优”,将类别数目N的更新进而导致的模型重建称为“模型变更”。传统的基于确定主动学习的多分类方法仅仅关注模型调优而忽略了模型变更,因此只适用于类别数目已知的应用场景;而在类别数目不确定的情况下,基于确定主动学习的多分类方法局限于现有N分类模型下样本信息量的评估,却无法准确描述和拟合样本数据的真实分布,从而无法实现分类性能的有效提升。For the sake of clarity, in the document of the present invention, the model parameters under the existing N classification model The optimization of is called "model tuning", and the model reconstruction caused by updating the number of categories N is called "model change". The traditional multi-classification method based on deterministic active learning only focuses on model tuning and ignores model changes, so it is only suitable for application scenarios where the number of categories is known; while the number of categories is uncertain, the multi-classification method based on deterministic active learning The method is limited to the evaluation of the amount of sample information under the existing N classification model, but it cannot accurately describe and fit the real distribution of the sample data, so that it cannot effectively improve the classification performance.

发明内容Contents of the invention

本发明的目的在于提供一种基于非确定主动学习的多分类器训练方法及分类方法,一方面对样本在现有模型框架下优化模型参数的能力进行评估,另一方面对该样本引入新的类别从而触发模型重建的可能性进行评估,通过综合考虑模型调优和模型变更两方面因素,实现样本信息量的综合评价,从而获得高效化、智能化的海量数据分类模型。The purpose of the present invention is to provide a multi-classifier training method and classification method based on non-deterministic active learning. On the one hand, it evaluates the ability of the sample to optimize model parameters under the existing model framework, and on the other hand, it introduces new Classes can be used to evaluate the possibility of triggering model reconstruction. By comprehensively considering the two factors of model optimization and model change, the comprehensive evaluation of sample information can be realized, so as to obtain an efficient and intelligent massive data classification model.

1、所提供的基于非确定主动学习的多分类方法,分别从模型变更和模型调优两个方面度量样本的信息量,一方面对样本在现有模型框架下优化模型参数的能力进行评估,另一方面对样本标注为新的类别从而触发模型重建的可能性进行评估,通过综合两方面因素,实现样本信息量的综合评价,按此评价标准选择信息量最大的样本进行标注,可以保证有限样本标注量下分类效果的最优化,从而获得高效化、智能化的海量数据分类模型。1. The provided multi-classification method based on non-deterministic active learning measures the amount of information of samples from two aspects of model change and model tuning, on the one hand, evaluates the ability of samples to optimize model parameters under the existing model framework, On the other hand, the possibility of triggering model reconstruction by labeling samples as new categories is evaluated. By combining two factors, a comprehensive evaluation of the amount of sample information is achieved. According to this evaluation standard, the sample with the largest amount of information is selected for labeling, which can ensure limited The optimization of the classification effect under the amount of sample labeling, so as to obtain an efficient and intelligent massive data classification model.

2、所提供的样本信息量计算方法,不仅考虑样本被标注为现有各类别的概率,而且考虑样本被标注为新类别的概率,形成样本信息量的统一化、综合性计算方法。2. The provided sample information calculation method not only considers the probability of samples being labeled as existing categories, but also considers the probability of samples being labeled as new categories, forming a unified and comprehensive calculation method for sample information.

3、所提供的基于聚类的样本批量选取方法,在对未标注集中样本进行聚类的基础上,批量选取最有信息的样本集,在保证样本信息量的同时避免了信息冗余。3. The provided clustering-based sample batch selection method, on the basis of clustering the unmarked samples, selects the most informative sample set in batches, which avoids information redundancy while ensuring the amount of sample information.

本发明的技术方案为:Technical scheme of the present invention is:

一种基于非确定主动学习的多分类器训练方法,其步骤为:A multi-classifier training method based on non-deterministic active learning, the steps of which are:

1)选取或初始化一多分类器;对未标注样本集中的每一样本,利用该多分类器计算该样本的总体信息量Info;所述总体信息量为:模型变更信息量与模型调优信息量之和;1) Select or initialize a multi-classifier; for each sample in the unlabeled sample set, use the multi-classifier to calculate the overall information amount Info of the sample; the overall information amount is: model change information amount and model tuning information the sum of the quantities;

2)对该未标注样本集进行聚类,得到J个子类;2) Clustering the unlabeled sample set to obtain J subclasses;

3)从每个子类中选取总体信息量Info值最小的若干未标注样本;再从所选未标注样本中选取K个样本进行标注后加入到已标注样本集L;3) Select a number of unlabeled samples with the smallest overall information Info value from each subcategory; then select K samples from the selected unlabeled samples for labeling and add them to the labeled sample set L;

4)将更新后的已标注样本集L作为训练数据重新训练该多分类器。4) Retrain the multi-classifier with the updated labeled sample set L as training data.

一种基于非确定主动学习的多分类器分类方法,其步骤为:A multi-classifier classification method based on non-deterministic active learning, the steps of which are:

1)选取或初始化一多分类器;对未标注样本集中的每一样本,利用该多分类器计算该样本的总体信息量Info;所述总体信息量为:模型变更信息量与模型调优信息量之和;1) Select or initialize a multi-classifier; for each sample in the unlabeled sample set, use the multi-classifier to calculate the overall information amount Info of the sample; the overall information amount is: model change information amount and model tuning information the sum of the quantities;

2)对该未标注样本集进行聚类,得到J个子类;2) Clustering the unlabeled sample set to obtain J subclasses;

3)从每个子类中选取总体信息量Info值最小的若干未标注样本;再从所选样本中选取K个样本进行标注后加入到已标注样本集L;3) Select a number of unlabeled samples with the smallest overall information Info value from each subcategory; then select K samples from the selected samples for labeling and add them to the labeled sample set L;

4)将更新后的已标注集L作为训练数据重新训练该多分类器;4) Retrain the multi-classifier with the updated labeled set L as training data;

5)迭代执行步骤1)~4)设定次数;然后利用最终得到的多分类器对未标注集进行分类。5) Iteratively execute steps 1) to 4) for a set number of times; then use the finally obtained multi-classifier to classify the unlabeled set.

进一步的,所述模型变更信息量为:从该未标注样本集中选取一样本a并将该样本的标注类别设定为新类别;然后利用该多分类器计算去除该样本a后的该未标注样本集关于该新类别的信息熵,将该信息熵作为该样本a的模型变更信息量;所述模型调优信息量的计算方法为:从该未标注样本集中选取一样本a并将该样本的标注类别设定为该多分类器中的一个类别;然后利用更新后的该多分类器计算去除该样本a的该未标注样本集关于每个已有类别的信息熵加权和,作为该样本a的模型调优信息量。Further, the information amount of the model change is: select a sample a from the unlabeled sample set and set the labeled category of the sample as a new category; then use the multi-classifier to calculate the unlabeled sample a after removing the sample a For the information entropy of the new category in the sample set, use the information entropy as the model change information amount of the sample a; the calculation method of the model tuning information amount is: select a sample a from the unlabeled sample set and add the sample a The labeled category of is set as a category in the multi-classifier; then the updated multi-classifier is used to calculate the weighted sum of the information entropy of each existing category in the unlabeled sample set that removes the sample a, as the sample The amount of model tuning information for a.

进一步的,计算所述模型变更信息量的方法为:首先根据具有N个类别训练数据的已标注样本集L构建一个N+1多分类器;然后对于去除该样本a后的该未标注样本集中每一样本x,将其不属于现有N个类别中任何一类的概率定义为该样本x属于第N+1个类别的概率;然后利用该多分类器计算去除该样本a后的该未标注样本集关于该新类别的信息熵,作为该样本a的模型变更信息量。Further, the method for calculating the amount of model change information is as follows: first construct an N+1 multi-classifier based on the labeled sample set L with N categories of training data; then for the unlabeled sample set after removing the sample a For each sample x, the probability that it does not belong to any of the existing N categories is defined as the probability that the sample x belongs to the N+1th category; Mark the information entropy of the new category in the sample set as the model change information amount of the sample a.

进一步的,计算所述模型变更信息量的方法为:首先根据具有N个类别训练数据的已标注样本集L构建一个二分类器,其中,将现有N个类别合并为一个类别A,将现有N个类别以外的其它类别归为另一类别B;然后对于去除该样本a后的该未标注样本集中每一样本x,将其不属于现有N个类别中任何一类的概率定义为该样本x属于类别B的概率;然后利用该多分类器计算去除该样本a后的该未标注样本集关于该新类别的信息熵,作为该样本a的模型变更信息量。Further, the method for calculating the amount of model change information is as follows: first construct a binary classifier based on the labeled sample set L with N categories of training data, wherein the existing N categories are merged into one category A, and the current There are other categories other than N categories classified into another category B; then, for each sample x in the unlabeled sample set after removing the sample a, the probability that it does not belong to any of the existing N categories is defined as The probability that the sample x belongs to category B; then use the multi-classifier to calculate the information entropy of the unlabeled sample set about the new category after removing the sample a, as the model change information amount of the sample a.

进一步的,计算所述模型变更信息量的方法为:首先根据具有N个类别训练数据的已标注样本集L构建一个一元分类器;然后对于去除该样本a后的该未标注样本集中每一样本x,将其不属于现有N个类别中任何一类的概率定义为样本x为离群点的概率;然后利用该多分类器计算去除该样本a后的该未标注样本集关于该新类别的信息熵,作为该样本a的模型变更信息量。Further, the method for calculating the amount of model change information is as follows: first construct a one-element classifier based on the labeled sample set L with N categories of training data; then for each sample in the unlabeled sample set after removing the sample a x, the probability that it does not belong to any of the existing N categories is defined as the probability that the sample x is an outlier; then use the multi-classifier to calculate the unlabeled sample set after removing the sample a about the new category The information entropy of is used as the model change information amount of the sample a.

本发明的主要内容包括:Main contents of the present invention include:

对于类别数目不确定的分类问题,类别数目N的取值为当前已标注集中样本不同标注的个数,随着已标注集的扩展,类别数目N随之调整。图1是类别数目不确定的分类模型构建过程示例:图1(1)中,初始的已标注集仅仅包含A、B两个已标注样本,分别属于类别1、类别2,因此相应的分类模型为二分类模型;图1(2)中,样本C被标注为类别1并加入已标注集,由于没有新标注加入,因此分类模型仍然为二分类模型;图1(3)中,样本D被标注为类别3并加入已标注集,由于新标注(类别3)的出现,分类模型变更为三分类模型。For classification problems with an uncertain number of categories, the value of the number of categories N is the number of different labels of samples in the current labeled set. With the expansion of the labeled set, the number of categories N will be adjusted accordingly. Figure 1 is an example of the construction process of a classification model with an uncertain number of categories: in Figure 1(1), the initial labeled set only contains two labeled samples A and B, which belong to category 1 and category 2 respectively, so the corresponding classification model It is a binary classification model; in Figure 1(2), sample C is marked as category 1 and added to the labeled set, since no new label is added, the classification model is still a binary classification model; in Figure 1(3), sample D is Labeled as category 3 and added to the labeled set, due to the emergence of new labels (category 3), the classification model was changed to a three-category model.

图1也表明了模型调优与模型变更这两个因素对于分类模型优化同样重要,不可偏废。图1(1)中,如果仅仅从现有二分类模型出发,样本C比样本D的信息量更大(因为样本C距离分类面更近,从而具有更大的不确定度);然而事实上,样本D对于分类模型的优化更有意义(因为样本D的标注不仅有助于模型参数更新,同时也引入了新标注信息,进而将模型重建为更加契合数据真实分布的三分类模型)。Figure 1 also shows that the two factors of model tuning and model change are equally important for classification model optimization and cannot be neglected. In Figure 1(1), if only starting from the existing binary classification model, sample C has more information than sample D (because sample C is closer to the classification surface, so it has greater uncertainty); however, in fact , sample D is more meaningful for the optimization of the classification model (because the annotation of sample D not only helps to update the model parameters, but also introduces new annotation information, and then rebuilds the model into a three-category model that better fits the real distribution of the data).

1、样本信息量度量1. Measurement of sample information

本发明提供的基于非确定主动学习的多分类方法将样本信息量有机融合到一个统一的框架下,实现样本信息量的综合有效度量。The multi-classification method based on non-deterministic active learning provided by the present invention organically integrates the amount of sample information into a unified framework, and realizes the comprehensive and effective measurement of the amount of sample information.

该方法从信息论出发,基于如下分析:(1)当一个样本被标注为新类别并加入已标注集,该样本向现有模型中引入了之前未曾建模的全新信息,从而增加了现有模型对于未标注样本的全局估计的不确定性,从信息论的角度,该样本会增加未标注集的总体信息熵;(2)当一个样本被标注为某个已知类别并加入已标注集,该样本为现有模型更好地拟合数据分布提供了新的约束条件,从信息论的角度,该样本往往会降低未标注集的总体信息熵。This method starts from information theory and is based on the following analysis: (1) When a sample is labeled as a new category and added to the labeled set, the sample introduces new information that has not been modeled before into the existing model, thereby increasing the existing model. For the uncertainty of the global estimate of an unlabeled sample, from the perspective of information theory, this sample will increase the overall information entropy of the unlabeled set; (2) when a sample is labeled as a known category and added to the labeled set, the The sample provides new constraints for the existing model to better fit the data distribution. From the perspective of information theory, the sample tends to reduce the overall information entropy of the unlabeled set.

基于上述分析,基于非确定主动学习的多分类方法从模型变更和模型调优两个方面度量样本的信息量:Based on the above analysis, the multi-classification method based on non-deterministic active learning measures the amount of sample information from two aspects: model change and model tuning:

(1)样本的“模型变更信息量”定义为:假设该样本被标注为新类别的情况下,利用该多分类器计算除去该样本的未标注集关于新类别的信息熵;公式化表示如下:(1) The "model change information amount" of a sample is defined as: assuming that the sample is labeled as a new category, use the multi-classifier to calculate the information entropy of the unlabeled set that removes the sample about the new category; the formula is as follows:

Info up grad e ( x ; φ L , θ L N ) = P ( y ∉ C N | x ; φ L ) F ( x , N + 1 ; θ L N + 1 ) .     公式(6) Info up grad e ( x ; φ L , θ L N ) = P ( the y ∉ C N | x ; φ L ) f ( x , N + 1 ; θ L N + 1 ) . Formula (6)

Infoupgrade与样本信息量正相关。公式中,为样本x在现有模型下被标注为新类别(即不属于现有N个类别中任何一类)的概率,φL为该概率模型的参数。Info upgrade is positively correlated with the amount of sample information. formula, is the probability that the sample x is marked as a new category (that is, does not belong to any of the existing N categories) under the existing model, and φ L is the parameter of the probability model.

(2)样本的“模型调优信息量”定义为:假设该样本被标注为已有N个类别中的一个,利用该多分类器计算除去该样本的未标注集关于每个已有类别的信息熵加权和;公式化表示如下:(2) The "model tuning information amount" of a sample is defined as: Assuming that the sample is marked as one of the existing N categories, the multi-classifier is used to calculate the unlabeled set of the sample except for each existing category The weighted sum of information entropy; the formula is as follows:

Info update ( x ; φ L , θ L N ) = P ( y ∈ C N | x ; φ L ) Σ y ~ ∈ C N P ( y ~ | x ; θ L N ) F ( x , y ~ ; θ L N ) = ( 1 - P ( y ∉ C N | x ; φ L ) ) Σ y ~ ∈ C N P ( y ~ | x ; θ L N ) F ( x , y ~ ; θ L N ) .     公式(7) Info update ( x ; φ L , θ L N ) = P ( the y ∈ C N | x ; φ L ) Σ the y ~ ∈ C N P ( the y ~ | x ; θ L N ) f ( x , the y ~ ; θ L N ) = ( 1 - P ( the y ∉ C N | x ; φ L ) ) Σ the y ~ ∈ C N P ( the y ~ | x ; θ L N ) f ( x , the y ~ ; θ L N ) . Formula (7)

Infoupdate与样本信息量负相关。Info update is negatively correlated with the amount of sample information.

基于非确定主动学习的多分类方法,将样本的模型变更信息量和模型调优信息量有机结合成为一个综合性度量;根据各自特点,重点监测Infoupgrade值显著高和Infoupdate值显著低的样本;公式化表示如下(但不限于该表示形式):Based on the multi-classification method of non-deterministic active learning, the amount of model change information and model tuning information of samples are organically combined into a comprehensive measure; according to their respective characteristics, focus on monitoring samples with significantly high Info upgrade values and significantly low Info update values ; is formulated as follows (but not limited to this representation):

x IMC - AL * = arg min x ∈ U Info ( x ; φ L , θ L N ) .     公式(8) x IMC - AL * = arg min x ∈ u Info ( x ; φ L , θ L N ) . Formula (8)

其中,in,

Info ( x ; φ L , θ L N ) = log [ Info update ( x ; φ L , θ L N ) = min x ∈ U Info update ( x ; φ L , θ L N ) + σ ] + λ log [ - ( Info up grad e ( x ; φ L , θ L N ) - max x ∈ U Info up grad e ( x ; φ L , θ L N ) ) + σ ] .     公式(9) Info ( x ; φ L , θ L N ) = log [ Info update ( x ; φ L , θ L N ) = min x ∈ u Info update ( x ; φ L , θ L N ) + σ ] + λ log [ - ( Info up grad e ( x ; φ L , θ L N ) - max x ∈ u Info up grad e ( x ; φ L , θ L N ) ) + σ ] . Formula (9)

Info为样本总信息量。公式中,λ是调整Infoupgrade和Infoupdate之间相对权重的参数,σ是为了避免计算结果出现-∞而人为加上的一个非常小的常量(如e-10),即根据公式(8)选择信息量最大的若干个样本进行标注。Info is the total information amount of the sample. In the formula, λ is a parameter to adjust the relative weight between Info upgrade and Info update , and σ is a very small constant (such as e -10 ) artificially added in order to avoid -∞ in the calculation result, that is, according to formula (8) Select several samples with the largest amount of information for labeling.

2、样本新类别概率计算2. Calculation of probability of new category of samples

在公式中,样本x不属于现有N个类别中任何一类的概率的计算有多种方法。给定现有已标注集L和N分类模型及其模型参数上述概率的计算方法包括但不限于以下三种:In the formula, the probability that sample x does not belong to any of the existing N classes There are many ways to calculate . Given the existing labeled sets L and N classification models and their model parameters The methods for calculating the above probability include but are not limited to the following three methods:

(1)基于已标注集L构建一个N+1分类模型,从而将样本x不属于现有N个类别中任何一类的概率定义为样本x属于第N+1个类别的概率。公式化表示如下:(1) Construct an N+1 classification model based on the labeled set L, so that the probability that the sample x does not belong to any of the existing N categories is defined as the probability that the sample x belongs to the N+1th category. Formulated as follows:

P ( y ∉ C N | x ; φ L ) = P ( y = N + 1 | x ; θ L N + 1 ) .     公式(10) P ( the y ∉ C N | x ; φ L ) = P ( the y = N + 1 | x ; θ L N + 1 ) . Formula (10)

(2)基于已标注集L构建一个二分类模型,将现有N个类别合并为一个类别“+1”,将现有N个类别以外的其它类别归为“-1”,从而将样本x不属于现有N个类别中任何一类的概率定义为样本x属于类别“-1”的概率。公式化表示如下:(2) Construct a binary classification model based on the labeled set L, merge the existing N categories into one category "+1", and classify other categories other than the existing N categories as "-1", so that the sample x The probability of not belonging to any of the existing N classes is defined as the probability that sample x belongs to class "-1". Formulated as follows:

P ( y ∉ C N | x ; φ L ) = P ( y = - 1 | x ; θ L 2 ) .     公式(11) P ( the y ∉ C N | x ; φ L ) = P ( the y = - 1 | x ; θ L 2 ) . Formula (11)

(3)基于已标注集L构建一元分类模型,该分类模型旨在从只包含一种类别的已标注集中训练分类器用以识别同类样本或检测离群样本,常用方法包括单类支持向量机(OC-SVM)。本方法将现有N个类别合并为一个类别,将OC-SVM输出值通过sigmoid函数转化为概率形式,从而将样本x不属于现有N个类别中任何一类的概率定义为样本x为离群点的概率。公式化表示如下:(3) Construct a one-element classification model based on the labeled set L. This classification model aims to train a classifier from a labeled set containing only one category to identify similar samples or detect outlier samples. Common methods include single-class support vector machines ( OC-SVM). This method merges the existing N categories into one category, and converts the output value of OC-SVM into a probability form through the sigmoid function, so that the probability that the sample x does not belong to any of the existing N categories is defined as the sample x is separated from Probability of cluster points. Formulated as follows:

P ( y ∉ C N | x ; φ L ) = 1 1 + exp ( - Ourput OC - SVM ( y = outlier | x ; L ) ) .     公式(12)。 P ( the y ∉ C N | x ; φ L ) = 1 1 + exp ( - Our put OC - SVM ( the y = outlier | x ; L ) ) . Formula (12).

3、基于聚类的样本批量选取机制3. Batch selection mechanism of samples based on clustering

实际应用中,为了保证方法执行效率,每次选出的最有信息样本不是一个而是一批(如K个样本)。如果仅仅根据公式选取Info值最小的K个样本,会不可避免地引入冗余信息,从而导致分类效率的下降。In practical applications, in order to ensure the execution efficiency of the method, the most informative samples selected each time are not one but a batch (such as K samples). If only the K samples with the smallest Info values are selected according to the formula, redundant information will inevitably be introduced, resulting in a decrease in classification efficiency.

本发明提供一种改进的样本批量选取方法——基于聚类的样本批量选取方法:(1)将未标注集中的样本聚为J(J≥K)类,聚类方法包括但不限于K-means、K-medoids、谱聚类等;(2)在每一类中,根据公式选取最有信息的样本,获得样本数为J的样本集;(3)在上述样本集中,根据公式选取最有信息(Info值最小)的K个样本。J是大于等于K的数用于处理信息冗余。The present invention provides an improved sample batch selection method—a cluster-based sample batch selection method: (1) cluster the samples in the unmarked set into J (J≥K) classes, and the clustering methods include but are not limited to K- means, K-medoids, spectral clustering, etc.; (2) in each category, select the most informative samples according to the formula, and obtain a sample set with the number of samples J; (3) in the above sample sets, select the most informative samples according to the formula K samples with information (minimum Info value). J is a number greater than or equal to K and is used to deal with information redundancy.

依据本发明方法选取K个最有信息样本之后:(1)对选取出的K个样本进行人工标注;(2)将标注后的K个样本从未标注集中去除,并加入已标注集;(3)基于新的已标注集,根据公式(3)训练新的分类模型,从而获得分类结果。After selecting the K most informative samples according to the method of the present invention: (1) manually label the selected K samples; (2) remove the marked K samples from the unmarked set and add to the marked set; ( 3) Based on the new labeled set, train a new classification model according to formula (3) to obtain classification results.

与现有技术相比,本发明的积极效果为:Compared with prior art, positive effect of the present invention is:

本发明所提供的基于非确定主动学习的多分类方法,分别从模型变更和模型调优两个方面度量样本的信息量,一方面对样本在现有模型框架下优化模型参数的能力进行评估,另一方面对该样本引入新的类别从而触发模型重建的可能性进行评估,通过综合两方面因素,实现样本信息量的综合、全面评价,从而为高效利用有限的标注样本获得最优化的海量数据分类结果提供了一种智能化解决方案。The multi-classification method based on non-deterministic active learning provided by the present invention measures the amount of information of samples from two aspects of model modification and model optimization, on the one hand, evaluates the ability of samples to optimize model parameters under the existing model framework, On the other hand, the possibility of introducing a new category into the sample to trigger model reconstruction is evaluated. By combining two factors, the comprehensive and comprehensive evaluation of the sample information is realized, so as to obtain the most optimized massive data for the efficient use of limited labeled samples. Classification results provide an intelligent solution.

附图说明Description of drawings

图1为类别数目不确定的分类模型构建过程示例;其中,Figure 1 is an example of the construction process of a classification model with an uncertain number of categories; where,

(1)初始的已标注集仅仅包含A、B两个已标注样本,(1) The initial labeled set only contains two labeled samples, A and B,

(2)样本C被标注为类别1并加入已标注集,(2) Sample C is labeled as category 1 and added to the labeled set,

(3)样本D被标注为类别3并加入已标注集;(3) Sample D is marked as category 3 and added to the marked set;

图2为本发明提供的基于非确定主动学习的多分类方法流程图。Fig. 2 is a flow chart of the multi-classification method based on non-deterministic active learning provided by the present invention.

具体实施方式Detailed ways

以下结合附图对本发明的原理和特征进行描述,所举实例只用于解释本发明,并非用于限定本发明的范围。The principles and features of the present invention are described below in conjunction with the accompanying drawings, and the examples given are only used to explain the present invention, and are not intended to limit the scope of the present invention.

实例基于非确定主动学习的多分类方法Example based non-deterministic active learning method for multi-classification

本发明提供的基于非确定主动学习的多分类方法通过循环迭代过程实现分类模型的逐步优化。The multi-classification method based on non-deterministic active learning provided by the invention realizes the gradual optimization of the classification model through a cyclic iterative process.

设每轮循环迭代需要标注K个样本,在每轮循环迭代内部执行以下流程:Assuming that each round of loop iteration needs to label K samples, the following process is executed inside each round of loop iteration:

方法执行结束后,设循环迭代次数为M,则通过人机交互由专家标注的样本总量为K×M。After the method is executed, if the number of loop iterations is M, the total number of samples marked by experts through human-computer interaction is K×M.

以图像分类为例,图像样本用颜色直方图、小波纹理等组成的特征向量x表示;初始已标注集中图像包括汽车、轮船、飞机、老虎、大象共5类,分别用数字1~5表示,则图像标注用y={1,2,…,5}表示;未标注图像构成未标注集U,已标注图像构成已标注集L表示;分类模型用后验概率P(y|x;θL)表示。Taking image classification as an example, the image sample is represented by a feature vector x composed of color histogram, wavelet texture, etc.; the initial set of marked images includes 5 categories of cars, ships, airplanes, tigers, and elephants, which are represented by numbers 1 to 5 , the image annotation is represented by y={1,2,...,5}; the unlabeled image constitutes the unlabeled set U, and the labeled image constitutes the labeled set L; the classification model uses the posterior probability P(y|x; θ L ) said.

为了提升分类模型的性能,需要选择一些未标注图像进行标注,并利用新的标注信息更新现有模型,假设每次模型更新需要新标注K=5个图像样本,迭代执行如下流程:In order to improve the performance of the classification model, it is necessary to select some unlabeled images for labeling, and update the existing model with new labeling information. Assuming that each model update requires new labeling K=5 image samples, iteratively execute the following process:

1)计算未标注图像样本的总体信息量Info(即该样本的模型变更信息量与模型调优信息量之和);1) Calculate the overall information amount Info of the unlabeled image sample (that is, the sum of the model change information amount and the model tuning information amount of the sample);

2)将未标注图像聚为J=10个子类;从每个子类中选择Info值最小的一个图像样本,共得到10个图像样本;在选出的10个图像样本中,选择Info值最小的5个图像样本;2) Cluster unlabeled images into J=10 subcategories; select an image sample with the smallest Info value from each subcategory, and obtain 10 image samples in total; among the selected 10 image samples, select the image sample with the smallest Info value 5 image samples;

3)标注选出的5个图像样本,并加入已标注集;3) Label the selected 5 image samples and add them to the labeled set;

4)将新的已标注集L作为训练数据重新训练图像分类模型;4) Retraining the image classification model with the new labeled set L as training data;

5)用更新后的分类模型对未标注集进行分类,进而获得改进的图像分类结果。5) Use the updated classification model to classify the unlabeled set, and then obtain improved image classification results.

本发明提供的基于非确定主动学习的多分类方法能够在类别数目不确定的情况下,利用有限的样本标注量获得最优化的分类效果。The multi-classification method based on the non-deterministic active learning provided by the present invention can obtain the optimal classification effect with a limited number of sample labels under the condition that the number of categories is uncertain.

Claims (10)

1. A multi-classifier training method based on non-deterministic active learning comprises the following steps:
1) selecting or initializing a multi-classifier; for each sample in the unlabeled sample set, calculating the total information amount Info of the sample by using the multi-classifier; the total information amount is: the sum of the model change information quantity and the model tuning information quantity;
2) clustering the unlabeled sample set to obtain J subclasses;
3) selecting a plurality of unlabeled samples with the minimum total information quantity Info value from each subclass; selecting K samples from the selected unlabeled samples, labeling the K samples, and adding the K samples into the labeled sample set L;
4) and using the updated labeled sample set L as training data to retrain the multi-classifier.
2. The method of claim 1, wherein the amount of model change information is: selecting a sample a from the unmarked sample set and setting the marked type of the sample as a new type; then, calculating the information entropy of the unlabeled sample set about the new class after the sample a is removed by using the multi-classifier, and taking the information entropy as the model change information quantity of the sample a; the calculation method of the model tuning information quantity comprises the following steps: selecting a sample a from the unlabeled sample set and setting the labeled class of the sample as one class of the multiple classifiers; and then, calculating the information entropy weighted sum of the unlabeled sample set without the sample a about each existing class by using the updated multi-classifier, and using the information entropy weighted sum as the model tuning information quantity of the sample a.
3. The method of claim 2, wherein the amount of model change information is calculated by: firstly, constructing an N +1 multi-classifier according to a labeled sample set L with N classes of training data; then, for each sample x in the unlabeled sample set after the sample a is removed, defining the probability that the sample x does not belong to any one of the existing N classes as the probability that the sample x belongs to the N +1 th class; and then, calculating the information entropy of the unlabeled sample set with the sample a removed with respect to the new class by using the multi-classifier, and taking the information entropy as the model change information quantity of the sample a.
4. The method of claim 2, wherein the amount of model change information is calculated by: firstly, constructing a two-classifier according to a labeled sample set L with N classes of training data, wherein the existing N classes are merged into a class A, and other classes except the existing N classes are classified into another class B; then, for each sample x in the unlabeled sample set after the sample a is removed, defining the probability that the sample x does not belong to any one of the existing N classes as the probability that the sample x belongs to the class B; and then, calculating the information entropy of the unlabeled sample set with the sample a removed with respect to the new class by using the multi-classifier, and taking the information entropy as the model change information quantity of the sample a.
5. The method of claim 2, wherein the amount of model change information is calculated by: firstly, constructing a unitary classifier according to a labeled sample set L with N types of training data; then, for each sample x in the unlabeled sample set after the sample a is removed, defining the probability that the sample x does not belong to any one of the existing N classes as the probability that the sample x is an outlier; and then, calculating the information entropy of the unlabeled sample set with the sample a removed with respect to the new class by using the multi-classifier, and taking the information entropy as the model change information quantity of the sample a.
6. A multi-classifier classification method based on non-deterministic active learning comprises the following steps:
1) selecting or initializing a multi-classifier; for each sample in the unlabeled sample set, calculating the total information amount Info of the sample by using the multi-classifier; the total information amount is: the sum of the model change information quantity and the model tuning information quantity;
2) clustering the unlabeled sample set to obtain J subclasses;
3) selecting a plurality of unlabeled samples with the minimum total information quantity Info value from each subclass; selecting K samples from the selected samples, labeling the K samples, and adding the K samples into a labeled sample set L;
4) retraining the multi-classifier by using the updated labeled set L as training data;
5) iteratively executing the steps 1) to 4) for a set number of times; and then classifying the unlabeled set by using the finally obtained multi-classifier.
7. The method of claim 6, wherein the amount of model change information is: selecting a sample a from the unmarked sample set and setting the marked type of the sample as a new type; then, calculating the information entropy of the unlabeled sample set about the new class after the sample a is removed by using the multi-classifier, and taking the information entropy as the model change information quantity of the sample a; the calculation method of the model tuning information quantity comprises the following steps: selecting a sample a from the unlabeled sample set and setting the labeled class of the sample as one class of the multiple classifiers; and then, calculating the information entropy weighted sum of the unlabeled sample set without the sample a about each existing class by using the updated multi-classifier, and using the information entropy weighted sum as the model tuning information quantity of the sample a.
8. The method of claim 7, wherein the amount of model change information is calculated by: firstly, constructing an N +1 multi-classifier according to a labeled sample set L with N classes of training data; then, for each sample x in the unlabeled sample set after the sample a is removed, defining the probability that the sample x does not belong to any one of the existing N classes as the probability that the sample x belongs to the N +1 th class; and then, calculating the information entropy of the unlabeled sample set with the sample a removed with respect to the new class by using the multi-classifier, and taking the information entropy as the model change information quantity of the sample a.
9. The method of claim 7, wherein the amount of model change information is calculated by: firstly, constructing a two-classifier according to a labeled sample set L with N classes of training data, wherein the existing N classes are merged into a class A, and other classes except the existing N classes are classified into another class B; then, for each sample x in the unlabeled sample set after the sample a is removed, defining the probability that the sample x does not belong to any one of the existing N classes as the probability that the sample x belongs to the class B; and then, calculating the information entropy of the unlabeled sample set with the sample a removed with respect to the new class by using the multi-classifier, and taking the information entropy as the model change information quantity of the sample a.
10. The method of claim 7, wherein the amount of model change information is calculated by: firstly, constructing a unitary classifier according to a labeled sample set L with N types of training data; then, for each sample x in the unlabeled sample set after the sample a is removed, defining the probability that the sample x does not belong to any one of the existing N classes as the probability that the sample x is an outlier; and then, calculating the information entropy of the unlabeled sample set with the sample a removed with respect to the new class by using the multi-classifier, and taking the information entropy as the model change information quantity of the sample a.
CN201510046879.8A 2015-01-29 2015-01-29 A kind of multi-categorizer training method and sorting technique based on non-determined Active Learning Active CN104657744B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510046879.8A CN104657744B (en) 2015-01-29 2015-01-29 A kind of multi-categorizer training method and sorting technique based on non-determined Active Learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510046879.8A CN104657744B (en) 2015-01-29 2015-01-29 A kind of multi-categorizer training method and sorting technique based on non-determined Active Learning

Publications (2)

Publication Number Publication Date
CN104657744A true CN104657744A (en) 2015-05-27
CN104657744B CN104657744B (en) 2017-10-24

Family

ID=53248843

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510046879.8A Active CN104657744B (en) 2015-01-29 2015-01-29 A kind of multi-categorizer training method and sorting technique based on non-determined Active Learning

Country Status (1)

Country Link
CN (1) CN104657744B (en)

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106294490A (en) * 2015-06-08 2017-01-04 富士通株式会社 The feature Enhancement Method of data sample and device and classifier training method and apparatus
CN106776925A (en) * 2016-11-30 2017-05-31 腾云天宇科技(北京)有限公司 A kind of Forecasting Methodology of mobile terminal user's sex, server and system
CN107004141A (en) * 2017-03-03 2017-08-01 香港应用科技研究院有限公司 Efficient labeling of large sample groups
CN107067025A (en) * 2017-02-15 2017-08-18 重庆邮电大学 A kind of data automatic marking method based on Active Learning
CN107545274A (en) * 2017-07-18 2018-01-05 北京建筑大学 Semi-supervised label ratio learning method
CN108108743A (en) * 2016-11-24 2018-06-01 百度在线网络技术(北京)有限公司 Abnormal user recognition methods and the device for identifying abnormal user
CN108596266A (en) * 2018-05-02 2018-09-28 深圳市易成自动驾驶技术有限公司 Blending decision method, device based on semi-supervised learning and storage medium
CN108875768A (en) * 2018-01-23 2018-11-23 北京迈格威科技有限公司 Data labeling method, device and system, and storage medium
CN109478229A (en) * 2016-08-31 2019-03-15 富士通株式会社 Training device, character recognition device and the method for sorter network for character recognition
CN109886925A (en) * 2019-01-19 2019-06-14 天津大学 An aluminum surface defect detection method combining active learning and deep learning
CN109934354A (en) * 2019-03-12 2019-06-25 北京信息科技大学 An abnormal data detection method based on active learning
CN109947989A (en) * 2019-03-18 2019-06-28 北京字节跳动网络技术有限公司 Method and apparatus for handling video
CN110222785A (en) * 2019-06-13 2019-09-10 重庆大学 Self-adapting confidence degree Active Learning Method for gas sensor drift correction
CN111160469A (en) * 2019-12-30 2020-05-15 湖南大学 Active learning method of target detection system
CN111611392A (en) * 2020-06-23 2020-09-01 中国人民解放军国防科技大学 An educational resource citation analysis method, system and medium that integrates multi-feature and voting strategies
CN112149721A (en) * 2020-09-10 2020-12-29 南京大学 A target detection method based on active learning to reduce labeling requirements
CN112200245A (en) * 2020-10-10 2021-01-08 深圳市华付信息技术有限公司 A semi-supervised image classification method
CN112528030A (en) * 2021-02-09 2021-03-19 中关村科学城城市大脑股份有限公司 Semi-supervised learning method and system for text classification
CN112784818A (en) * 2021-03-03 2021-05-11 电子科技大学 Identification method based on grouping type active learning on optical remote sensing image
CN113435492A (en) * 2021-06-21 2021-09-24 华中科技大学 Power system dominant instability mode discrimination method based on active learning
CN114092735A (en) * 2020-08-24 2022-02-25 北京航空航天大学 A self-labeling method and system for instance object target detection level
CN114419363A (en) * 2021-12-23 2022-04-29 北京三快在线科技有限公司 Target classification model training method and device based on label-free sample data

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7756800B2 (en) * 2006-12-14 2010-07-13 Xerox Corporation Method for transforming data elements within a classification system based in part on input from a human annotator/expert
CN101853400A (en) * 2010-05-20 2010-10-06 武汉大学 Multi-Class Image Classification Method Based on Active Learning and Semi-Supervised Learning
CN102176701A (en) * 2011-02-18 2011-09-07 哈尔滨工业大学 Active learning based network data anomaly detection method
CN103258214A (en) * 2013-04-26 2013-08-21 南京信息工程大学 Remote sensing image classification method based on image block active learning
CN103400144A (en) * 2013-07-17 2013-11-20 山东师范大学 Active learning method based on K-neighbor for support vector machine (SVM)
CN103955462A (en) * 2014-03-21 2014-07-30 南京邮电大学 Image marking method based on multi-view and semi-supervised learning mechanism

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7756800B2 (en) * 2006-12-14 2010-07-13 Xerox Corporation Method for transforming data elements within a classification system based in part on input from a human annotator/expert
CN101853400A (en) * 2010-05-20 2010-10-06 武汉大学 Multi-Class Image Classification Method Based on Active Learning and Semi-Supervised Learning
CN102176701A (en) * 2011-02-18 2011-09-07 哈尔滨工业大学 Active learning based network data anomaly detection method
CN103258214A (en) * 2013-04-26 2013-08-21 南京信息工程大学 Remote sensing image classification method based on image block active learning
CN103400144A (en) * 2013-07-17 2013-11-20 山东师范大学 Active learning method based on K-neighbor for support vector machine (SVM)
CN103955462A (en) * 2014-03-21 2014-07-30 南京邮电大学 Image marking method based on multi-view and semi-supervised learning mechanism

Cited By (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106294490A (en) * 2015-06-08 2017-01-04 富士通株式会社 The feature Enhancement Method of data sample and device and classifier training method and apparatus
CN106294490B (en) * 2015-06-08 2019-12-24 富士通株式会社 Feature enhancement method and device for data samples and classifier training method and device
CN109478229A (en) * 2016-08-31 2019-03-15 富士通株式会社 Training device, character recognition device and the method for sorter network for character recognition
CN109478229B (en) * 2016-08-31 2021-08-10 富士通株式会社 Training device for classification network for character recognition, character recognition device and method
CN108108743B (en) * 2016-11-24 2022-06-24 百度在线网络技术(北京)有限公司 Abnormal user identification method and device for identifying abnormal user
CN108108743A (en) * 2016-11-24 2018-06-01 百度在线网络技术(北京)有限公司 Abnormal user recognition methods and the device for identifying abnormal user
CN106776925B (en) * 2016-11-30 2020-07-14 腾云天宇科技(北京)有限公司 Method, server and system for predicting gender of mobile terminal user
CN106776925A (en) * 2016-11-30 2017-05-31 腾云天宇科技(北京)有限公司 A kind of Forecasting Methodology of mobile terminal user's sex, server and system
CN107067025A (en) * 2017-02-15 2017-08-18 重庆邮电大学 A kind of data automatic marking method based on Active Learning
CN107067025B (en) * 2017-02-15 2020-12-22 重庆邮电大学 An automatic labeling method for text data based on active learning
CN107004141A (en) * 2017-03-03 2017-08-01 香港应用科技研究院有限公司 Efficient labeling of large sample groups
CN107545274A (en) * 2017-07-18 2018-01-05 北京建筑大学 Semi-supervised label ratio learning method
CN108875768A (en) * 2018-01-23 2018-11-23 北京迈格威科技有限公司 Data labeling method, device and system, and storage medium
CN108596266A (en) * 2018-05-02 2018-09-28 深圳市易成自动驾驶技术有限公司 Blending decision method, device based on semi-supervised learning and storage medium
CN109886925A (en) * 2019-01-19 2019-06-14 天津大学 An aluminum surface defect detection method combining active learning and deep learning
CN109934354A (en) * 2019-03-12 2019-06-25 北京信息科技大学 An abnormal data detection method based on active learning
CN109947989A (en) * 2019-03-18 2019-06-28 北京字节跳动网络技术有限公司 Method and apparatus for handling video
CN109947989B (en) * 2019-03-18 2023-08-29 北京字节跳动网络技术有限公司 Method and apparatus for processing video
CN110222785A (en) * 2019-06-13 2019-09-10 重庆大学 Self-adapting confidence degree Active Learning Method for gas sensor drift correction
CN110222785B (en) * 2019-06-13 2023-10-13 重庆大学 Adaptive confidence active learning method for gas sensor drift correction
CN111160469A (en) * 2019-12-30 2020-05-15 湖南大学 Active learning method of target detection system
CN111160469B (en) * 2019-12-30 2023-08-08 湖南大学 An Active Learning Method for Object Detection System
CN111611392A (en) * 2020-06-23 2020-09-01 中国人民解放军国防科技大学 An educational resource citation analysis method, system and medium that integrates multi-feature and voting strategies
CN114092735A (en) * 2020-08-24 2022-02-25 北京航空航天大学 A self-labeling method and system for instance object target detection level
CN114092735B (en) * 2020-08-24 2024-11-01 北京航空航天大学 Self-labeling method and system for object detection level of instance object
CN112149721A (en) * 2020-09-10 2020-12-29 南京大学 A target detection method based on active learning to reduce labeling requirements
CN112149721B (en) * 2020-09-10 2023-11-17 南京大学 A target detection method based on active learning to reduce labeling requirements
CN112200245A (en) * 2020-10-10 2021-01-08 深圳市华付信息技术有限公司 A semi-supervised image classification method
CN112528030A (en) * 2021-02-09 2021-03-19 中关村科学城城市大脑股份有限公司 Semi-supervised learning method and system for text classification
CN112784818A (en) * 2021-03-03 2021-05-11 电子科技大学 Identification method based on grouping type active learning on optical remote sensing image
CN112784818B (en) * 2021-03-03 2023-03-14 电子科技大学 Identification method based on grouping type active learning on optical remote sensing image
CN113435492A (en) * 2021-06-21 2021-09-24 华中科技大学 Power system dominant instability mode discrimination method based on active learning
CN113435492B (en) * 2021-06-21 2022-04-22 华中科技大学 Power system dominant instability mode discrimination method based on active learning
CN114419363A (en) * 2021-12-23 2022-04-29 北京三快在线科技有限公司 Target classification model training method and device based on label-free sample data

Also Published As

Publication number Publication date
CN104657744B (en) 2017-10-24

Similar Documents

Publication Publication Date Title
CN104657744B (en) A kind of multi-categorizer training method and sorting technique based on non-determined Active Learning
Nam et al. Transfer defect learning
Ramu et al. A survey of machine learning techniques in structural and multidisciplinary optimization
Singer et al. Efficient learning using forward-backward splitting
Li et al. Learning balanced and unbalanced graphs via low-rank coding
CN113569906A (en) Method and Device for Extracting Heterogeneous Graph Information Based on Meta-Path Subgraph
Wang et al. Deep generative model for robust imbalance classification
Fisch et al. Conformal prediction sets with limited false positives
Liu et al. Hierarchical adaptive pooling by capturing high-order dependency for graph representation learning
CN105005794A (en) Image pixel semantic annotation method with combination of multi-granularity context information
Zhu et al. Multiview latent space learning with progressively fine-tuned deep features for unsupervised domain adaptation
CN110516950A (en) A Risk Analysis Method Oriented to Entity Resolution Task
CN106777006A (en) A kind of sorting algorithm based on parallel super-network under Spark
Liu et al. MI-ELM: highly efficient multi-instance learning based on hierarchical extreme learning machine
Kalifullah et al. Retracted: Graph‐based content matching for web of things through heuristic boost algorithm
Fadhil Hybrid of K-means clustering and naive Bayes classifier for predicting performance of an employee
Behrouz et al. Chimera: Effectively modeling multivariate time series with 2-dimensional state space models
Grishma et al. Software root cause prediction using clustering techniques: A review
Jung et al. Learning to solve minimum cost multicuts efficiently using edge-weighted graph convolutional neural networks
CN108664607A (en) A kind of power telecom network quality of data method for improving based on transfer learning
Banjongkan et al. A comparative study of learning techniques with convolutional neural network based on HPC-workload dataset
Liu et al. Meta hyperparameter optimization with adversarial proxy subsets sampling
Ying et al. Topology-aware Reinforcement Feature Space Reconstruction for Graph Data
Tao et al. Resilient learning of computational models with noisy labels
Li et al. ipath: Forecasting the pathway to impact

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant