[go: up one dir, main page]

CN117079024A - Image class increment learning algorithm integrating uncertainty estimation and increment stage discrimination - Google Patents

Image class increment learning algorithm integrating uncertainty estimation and increment stage discrimination Download PDF

Info

Publication number
CN117079024A
CN117079024A CN202311034343.5A CN202311034343A CN117079024A CN 117079024 A CN117079024 A CN 117079024A CN 202311034343 A CN202311034343 A CN 202311034343A CN 117079024 A CN117079024 A CN 117079024A
Authority
CN
China
Prior art keywords
classification
stage
increment
incremental
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311034343.5A
Other languages
Chinese (zh)
Inventor
曾明
李思颖
钟舒桐
赵峰
王湘晖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN202311034343.5A priority Critical patent/CN117079024A/en
Publication of CN117079024A publication Critical patent/CN117079024A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Multimedia (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

The application relates to an image class increment learning algorithm integrating uncertainty estimation and increment stage discrimination, which comprises the following steps: a large-scale pre-training network (such as a CLIP network) is adopted as a characteristic extraction network of the classification model, parameters of the characteristic extraction network are frozen in the training process, a classification head is independently trained for each incremental stage, and meanwhile, the classification head trained before is reserved. And in the classification, the category with the highest confidence in the classification head with the smallest classification uncertainty is used as the classification result of the image. According to the application, the parameters of the feature extraction network are shared in each incremental stage, so that the parameter calculation amount of the model is greatly reduced. The large-scale pre-training network is used as the characteristic extraction network of the classification model, so that the classification model has good generalization in all incremental stages, has good discrimination on images of new and old categories, and greatly improves the classification accuracy of the incremental learning model.

Description

融合不确定度估计和增量阶段判别的图像类增量学习算法Image class incremental learning algorithm integrating uncertainty estimation and incremental stage discrimination

技术领域Technical field

本发明涉及图像类增量学习领域,具体来说,是一种融合分类不确定度估计和增量阶段判别的图像类增量学习算法。The invention relates to the field of image class incremental learning. Specifically, it is an image class incremental learning algorithm that integrates classification uncertainty estimation and incremental stage discrimination.

背景技术Background technique

人类和动物能够应对环境的不断变化接受和学习新的知识,并对原有的知识进行补充和修正。在很多分类任务中,也需要分类算法具备人类和动物这种持续学习新知识,并避免遗忘旧知识的能力,这类算法称为增量学习算法。本发明研究图像类增量学习问题,即分类模型在保持对旧类别的判别力同时,可持续学习新的类别。目前主流的图像分类算法是在静态环境中识别有限类别数量的图片,但是在实际应用过程中会不断的出现新的类别,大部分神经网络并不能识别未训练过的图片。即神经网络模型仅能识别已知类别的图片,对于未训练的图片种类,神经网络模型会将该物体错误的归到已知的类别,从而导致识别精度下降,所以需要设计图像增量学习算法,使得模型可以在保持对旧类别的判别力同时,可持续学习新类别的知识。Humans and animals can accept and learn new knowledge in response to constant changes in the environment, and supplement and modify the original knowledge. In many classification tasks, classification algorithms also need to have the ability of humans and animals to continuously learn new knowledge and avoid forgetting old knowledge. This type of algorithm is called an incremental learning algorithm. The present invention studies the problem of incremental learning of image classes, that is, the classification model can continuously learn new categories while maintaining the discriminative power of old categories. The current mainstream image classification algorithm is to identify a limited number of images in a static environment. However, new categories will continue to appear in the actual application process, and most neural networks cannot recognize untrained images. That is, the neural network model can only recognize pictures of known categories. For untrained picture types, the neural network model will mistakenly classify the object into a known category, resulting in a decrease in recognition accuracy. Therefore, it is necessary to design an image incremental learning algorithm. , so that the model can continuously learn the knowledge of new categories while maintaining the discriminative power of old categories.

增量学习分类模型一方面必须表现出从新数据中学习新知识和提炼已有知识的能力,另一方面又必须避免新知识对已有知识的显著干扰。如何让模型更好的在两者间取得平衡,是增量学习算法所面临的挑战。传统的类增量学习算法都是从头开始训练分类模型的,以便在每次新类别的数据到来时不断地获取到新的知识。目前已经提出了一些类别增量学习策略,其中主流的方法是在从头训练的基类模型上,存储少量的旧类别样本与新增的类别样本一起联合微调模型参数,但是这类算法会使得模型的参数不断变化,容易导致模型遗忘旧类别的知识。最近大规模预训练模型在深度学习领域取得重大突破,它使用了海量的数据集进行训练,大规模预训练模型强大的特征提取能力和泛化性使得它可以很好地应用在增量学习算法中。On the one hand, the incremental learning classification model must show the ability to learn new knowledge from new data and refine existing knowledge, and on the other hand, it must avoid significant interference of new knowledge on existing knowledge. How to make the model better balance between the two is a challenge faced by incremental learning algorithms. Traditional incremental learning algorithms train classification models from scratch in order to continuously acquire new knowledge every time new categories of data arrive. Some category incremental learning strategies have been proposed. The mainstream method is to store a small number of old category samples and jointly fine-tune the model parameters with the new category samples on the base class model trained from scratch. However, this type of algorithm will make the model The parameters are constantly changing, which can easily cause the model to forget the knowledge of old categories. Recently, large-scale pre-training models have made major breakthroughs in the field of deep learning. They use massive data sets for training. The powerful feature extraction capabilities and generalization capabilities of large-scale pre-training models make them well suited for incremental learning algorithms. middle.

本发明提供了一种融合不确定度估计和增量阶段判别的图像类增量学习算法,采用大规模预训练网络作为分类模型的特征提取网络,在训练过程中冻结特征提取网络的参数以保证特征提取网络对增量过程中每一阶段类别的泛化性,同时为了能够区分每一个阶段的图像类别,为每一个阶段单独训练一个分类头。在分类时,先计算图像在各个分类头上的分类不确定度确定出该图像所属的增量阶段,将分类不确定度最小的分类头所属的阶段作为该图像所属的增量阶段,然后选择该增量阶段对应的分类头中置信度最高的类别作为该图像的分类结果。The present invention provides an image class incremental learning algorithm that integrates uncertainty estimation and incremental stage discrimination, uses a large-scale pre-training network as the feature extraction network of the classification model, and freezes the parameters of the feature extraction network during the training process to ensure The feature extraction network generalizes to the categories at each stage in the incremental process. At the same time, in order to be able to distinguish the image categories at each stage, a separate classification head is trained for each stage. When classifying, first calculate the classification uncertainty of the image on each classification head to determine the incremental stage to which the image belongs. The stage to which the classification head with the smallest classification uncertainty belongs is regarded as the incremental stage to which the image belongs, and then select The category with the highest confidence in the classification head corresponding to this incremental stage is used as the classification result of the image.

发明内容Contents of the invention

针对在增量过程中不断学习新数据的知识同时不忘记旧数据知识的迫切需求,本发明提供了一种融合不确定度估计和增量阶段判别的图像类增量学习算法。本发明采用如下技术方案:In response to the urgent need to continuously learn the knowledge of new data while not forgetting the knowledge of old data during the incremental process, the present invention provides an image class incremental learning algorithm that integrates uncertainty estimation and incremental stage discrimination. The present invention adopts the following technical solutions:

一种融合不确定度估计和增量阶段判别的图像类增量学习算法,包括以下步骤:An image class incremental learning algorithm that integrates uncertainty estimation and incremental stage discrimination, including the following steps:

1)将大规模预训练网络(例如CLIP网络)作为分类模型的特征提取网络,冻结特征提取网络的参数,将基类的数据输入所述特征提取网络,使用交叉熵损失函数LCE训练基类阶段的分类头参数,分类头的类别数量为基类数据的类别数量N01) Use a large-scale pre-trained network (such as CLIP network) as the feature extraction network of the classification model, freeze the parameters of the feature extraction network, input the data of the base class into the feature extraction network, and use the cross-entropy loss function L CE to train the base class The classification head parameters of the stage, the number of categories of the classification head is the number of categories of base class data N 0 ;

2)对于第k个增量阶段的训练,冻结特征提取网络的参数,并保留之前的增量阶段训练后的分类头,将新类别的数据输入特征提取网络,使用交叉熵损失函数L′CE训练增量阶段的分类头参数,分类头的类别数量为该阶段增量数据的类别数量Nk2) For the training of the kth incremental stage, freeze the parameters of the feature extraction network and retain the classification heads after training in the previous incremental stage. Enter the new category data into the feature extraction network and use the cross-entropy loss function L′ CE Classification head parameters in the incremental phase of training, the number of categories of the classification head is the number of categories N k of the incremental data in this phase;

3)在测试阶段,计算待分类图像x在第k个增量阶段的分类头上的分类不确定度Hk(x),将分类不确定度最小的分类头所属的增量阶段作为该图像所属的增量阶段;3) In the test phase, calculate the classification uncertainty H k (x) of the classification head of the k-th incremental stage of the image x to be classified, and assign the classification head with the smallest classification uncertainty to the incremental stage to which as the incremental stage to which this image belongs;

4)将第个增量阶段的分类头中分类置信度最高的类别作为图像x的分类结果 4) General The category with the highest classification confidence in the classification heads of the incremental stages is used as the classification result of image x

步骤1)中交叉熵损失函数LCE的计算公式为:The calculation formula of the cross entropy loss function L CE in step 1) is:

其中,yi表示模型分类头的输出,gi表示真实的标签值,N0为基类阶段的分类头的类别数量。Among them, yi represents the output of the model classification head, gi represents the real label value, and N 0 is the number of categories of the classification head in the base class stage.

步骤2)中交叉熵损失函数L′CE的计算公式为:The calculation formula of the cross entropy loss function L′ CE in step 2) is:

其中,yi表示模型分类头的输出,gi表示真实的标签值,Nk为第k个增量阶段的分类头的类别数量。Among them, yi represents the output of the model classification head, gi represents the real label value, and N k is the number of categories of the classification head in the kth incremental stage.

步骤3)中图像x在第k个分类头上的分类不确定度Hk(x)的计算公式为:The calculation formula for the classification uncertainty H k (x) of image x on the k-th classification head in step 3) is:

其中,yki(x)表示图像x在第k个增量阶段的分类头上的第i个输出,Nk表示第k个增量阶段的分类头的类别数量。Among them, y ki (x) represents the i-th output of the classification head of image x in the k-th incremental stage, and N k represents the number of categories of the classification head in the k-th incremental stage.

步骤3)中分类不确定度最小的分类头所属的增量阶段的计算公式为:The incremental stage to which the classification head with the smallest classification uncertainty belongs in step 3) The calculation formula is:

其中,n表示到目前为止增量阶段的阶段数量,增量阶段k取值为0时,表示该增量阶段是基类阶段,是增量学习模型训练的起点。Among them, n represents the number of stages in the incremental stage so far. When the value of incremental stage k is 0, it means that the incremental stage is the base class stage and is the starting point for incremental learning model training.

步骤4)中图像x的分类结果的计算公式为:Classification results of image x in step 4) The calculation formula is:

其中,表示图像x在增量阶段/>的分类头上的第i个输出,/>表示增量阶段/>的分类头的类别数量。in, Represents the image x in the incremental phase /> The i-th output of the classification head,/> Represents the incremental stage/> The number of categories in the category header.

本发明具有以下有益效果:1)使用大规模预训练网络作为分类模型的特征提取网络,使得分类模型在所有的增量阶段都保持良好的泛化性,并且特征提取网络具有强大的特征提取能力,对新类和旧类的图像都能保持很好的辨别力;2)为每个增量阶段的数据都单独训练一个分类头,并使用分类不确定度来确定待测试图像属于第几个增量阶段,从而判断应该使用的分类头,大幅提升增量学习模型的分类准确率。3)由于特征提取网络的参数被冻结,每一个增量阶段可以共享参数,因此可训练的参数大幅度减少。The present invention has the following beneficial effects: 1) Using a large-scale pre-training network as the feature extraction network of the classification model enables the classification model to maintain good generalization in all incremental stages, and the feature extraction network has powerful feature extraction capabilities , both new and old categories of images can maintain good discrimination; 2) A separate classification head is trained for the data in each incremental stage, and the classification uncertainty is used to determine which number the image to be tested belongs to. In the incremental stage, it can determine the classification head that should be used and greatly improve the classification accuracy of the incremental learning model. 3) Since the parameters of the feature extraction network are frozen, parameters can be shared in each incremental stage, so the trainable parameters are greatly reduced.

附图说明Description of the drawings

图1为本发明实施例中融合不确定度估计和增量阶段判别的图像类增量学习算法的流程示意图Figure 1 is a schematic flow chart of an image class incremental learning algorithm that combines uncertainty estimation and incremental stage discrimination in an embodiment of the present invention.

图2为本发明实施例中的方法与其他方法对比的结果Figure 2 shows the comparison results between the method in the embodiment of the present invention and other methods.

具体实施方式Detailed ways

下面结合附图和具体实施方式对发明作进一步的详细说明。实施例是对本发明中技术方案清楚完整的描述,所描述的实施例是本发明一部分实施例而不是全部实施例。此处所描述的具体实施例方式仅用于解释本发明,并不用于限定本发明的保护范围。The invention will be further described in detail below with reference to the accompanying drawings and specific embodiments. The embodiments are a clear and complete description of the technical solutions in the present invention. The described embodiments are part of the embodiments of the present invention but not all of them. The specific embodiments described here are only used to explain the present invention and are not used to limit the scope of the present invention.

1)本发明采用CIFAR-10数据集,该数据集是一个经典的用于图像识别和分类的数据集,包含10个类别、共计60000张32x32大小的彩色图片,其中训练集有50000张,测试集有10000张。本发明采用CLIP大规模预训练网络作为分类模型的特征提取网络,CLIP是一种基于对比学习的多模态模型,它使用了大规模的数据集去训练具有强大泛化能力的特征提取网络。本发明包括多个阶段的图像分类模型训练,但是为了更加清楚地说明本发明的相关技术特征,本发明以第零个增量阶段、第一个增量阶段和第二个增量阶段的增量学习模型训练作为示例进行介绍;1) This invention uses the CIFAR-10 data set, which is a classic data set used for image recognition and classification. It contains 10 categories and a total of 60,000 color pictures of 32x32 size, of which the training set has 50,000 pictures. The test set The set has 10,000 pictures. The present invention adopts CLIP large-scale pre-training network as the feature extraction network of the classification model. CLIP is a multi-modal model based on contrastive learning. It uses a large-scale data set to train a feature extraction network with strong generalization capabilities. The present invention includes multiple stages of image classification model training. However, in order to explain the relevant technical features of the present invention more clearly, the present invention uses the incremental stages of the zeroth incremental stage, the first incremental stage and the second incremental stage. Quantitative learning model training is introduced as an example;

2)第零个增量阶段的增量学习模型训练:该阶段训练的模型是基类模型,其为增量学习模型训练的起点,随机选取CIFAR-10数据集中的六个类别作为第零个增量阶段的基类数据集,基类模型的特征提取网络为CLIP预训练网络,在训练过程中冻结特征提取网络的参数,使用交叉熵损失LCE训练基类模型的分类头:2) Incremental learning model training in the zeroth incremental stage: The model trained in this stage is the base class model, which is the starting point for incremental learning model training. Six categories in the CIFAR-10 data set are randomly selected as the zeroth For the base class data set in the incremental stage, the feature extraction network of the base class model is a CLIP pre-training network. The parameters of the feature extraction network are frozen during the training process, and the cross-entropy loss L CE is used to train the classification head of the base class model:

其中,yi表示基类模型分类头的输出,gi表示基类数据集真实的标签值,N0为基类阶段的分类头的类别数量。Among them, yi represents the output of the classification head of the base class model, gi represents the real label value of the base class data set, and N 0 is the number of categories of the classification head in the base class stage.

3)第一个增量阶段的增量学习模型的训练:其为在第一个增量阶段的训练的基础上进行的类别增量学习训练。使用第零个增量阶段的特征提取网络作为第一个增量阶段的特征提取网络,同样冻结特征提取网络的参数,并保留第零个增量阶段训练后的分类头。随机选取CIFAR-10数据集中剩下的两个类别的数据作为第一个增量阶段的增量数据集,使用交叉熵损失L′CE训练第一个增量阶段的增量学习模型的分类头参数:3) Training of the incremental learning model in the first incremental stage: It is category incremental learning training based on the training in the first incremental stage. Use the feature extraction network of the zeroth incremental stage as the feature extraction network of the first incremental stage, also freeze the parameters of the feature extraction network, and retain the classification head trained in the zeroth incremental stage. Randomly select the remaining two categories of data in the CIFAR-10 data set as the incremental data set in the first incremental stage, and use the cross-entropy loss L′ CE to train the classification head of the incremental learning model in the first incremental stage. parameter:

其中,yi表示第一个增量阶段增量学习模型分类头的输出,gi表示第一个增量阶段增量数据集真实的标签值,Nk为第一个增量阶段的分类头的类别数量。Among them, yi represents the output of the classification head of the incremental learning model in the first incremental stage, gi represents the real label value of the incremental data set in the first incremental stage, and N k is the classification head of the first incremental stage. number of categories.

4)第二个增量阶段的增量学习模型的训练:其为在第一个增量阶段的训练的基础上进行的类别增量学习训练。使用第一个增量阶段的特征提取网络作为第二个增量阶段的特征提取网络,同样冻结特征提取网络的参数,并保留第零个增量阶段和第一个增量阶段训练后的分类头。选取CIFAR-10数据集中剩下的两个类别的数据作为第二个增量阶段的增量数据集,使用交叉熵损失L′CE训练第二个增量阶段的增量学习模型的分类头参数:4) Training of the incremental learning model in the second incremental stage: It is category incremental learning training based on the training in the first incremental stage. Use the feature extraction network in the first incremental stage as the feature extraction network in the second incremental stage, also freeze the parameters of the feature extraction network, and retain the classification after training in the zeroth incremental stage and the first incremental stage head. Select the remaining two categories of data in the CIFAR-10 data set as the incremental data set in the second incremental stage, and use the cross-entropy loss L′ CE to train the classification head parameters of the incremental learning model in the second incremental stage. :

其中,yi表示第二个增量阶段增量学习模型分类头的输出,gi表示第二个增量阶段增量数据集真实的标签值,Nk为第二个增量阶段的分类头的类别数量。Among them, yi represents the output of the classification head of the incremental learning model in the second incremental stage, gi represents the real label value of the incremental data set in the second incremental stage, and N k is the classification head of the second incremental stage. number of categories.

5)在测试阶段,计算待分类图像x在第k个增量阶段的分类头上的分类不确定度Hk(x),将分类不确定度最小的分类头所属的增量阶段作为该图像所属的增量阶段:5) In the test phase, calculate the classification uncertainty H k (x) of the classification head of the k-th incremental stage of the image x to be classified, and assign the classification head with the smallest classification uncertainty to the incremental stage to which As the incremental stage this image belongs to:

其中,yki(x)表示图像x在第k个增量阶段的分类头上的第i个输出,Nk表示第k个增量阶段的分类头的类别数量,n表示到目前为止增量阶段的阶段数量。Among them, y ki (x) represents the i-th output of the classification head of image x in the k-th incremental stage, N k represents the number of categories of the classification head in the k-th incremental stage, and n represents the increment so far. The number of stages for the stage.

6)将第个增量阶段的分类头中分类置信度最高的类别作为图像x的分类结果 6) General The category with the highest classification confidence in the classification heads of the incremental stages is used as the classification result of image x

其中,表示图像x在增量阶段/>的分类头上的第i个输出,/>表示增量阶段/>的分类头的类别数量。in, Represents the image x in the incremental phase /> The i-th output of the classification head,/> Represents the incremental stage/> The number of categories in the category header.

7)图2给出了使用传统的从头开始训练的增量学习方法PASS(为每一个旧类别保留一个原型,并在特征空间中采用基于原型的区域扩充来维持旧类别的分类边界)和本发明方法的识别准确率,可以看出,本发明方法显著提高了每个阶段模型的识别准确率,第零个增量阶段识别率提升了2.22%,第一个增量阶段识别率提升了12.55%,第二个增量阶段识别率提升了8.74%。7) Figure 2 shows the traditional incremental learning method PASS trained from scratch (retaining a prototype for each old category and using prototype-based region expansion in the feature space to maintain the classification boundaries of the old category) and this From the recognition accuracy of the inventive method, it can be seen that the inventive method significantly improves the recognition accuracy of the model in each stage. The recognition rate in the zeroth incremental stage is increased by 2.22%, and the recognition rate in the first incremental stage is increased by 12.55%. %, the recognition rate increased by 8.74% in the second incremental stage.

由此可见,本发明提出的一种融合不确定度估计和增量阶段判别的图像类增量学习算法,可以有效的提高增量学习识别准确率,为后续的增量学习算法研究以及工程应用提供新的思路。It can be seen that the image-based incremental learning algorithm proposed by the present invention that integrates uncertainty estimation and incremental stage discrimination can effectively improve the incremental learning recognition accuracy and provide a basis for subsequent incremental learning algorithm research and engineering applications. Provide new ideas.

以上,仅为本发明较佳的一种实施方案,并非用于限制本发明的保护范围,任何有关技术领域的普通技术人员在本申请揭露的技术范围内,进行等同变化或等效替换所获得的方案,都应涵盖在本申请的保护范围之内。权利要求书规定的保护范围为本申请的保护范围。The above is only a preferred embodiment of the present invention and is not used to limit the protection scope of the present invention. Any person of ordinary skill in the relevant technical field can make equivalent changes or equivalent substitutions within the technical scope disclosed in this application. All solutions shall be covered by the protection scope of this application. The protection scope specified in the claims is the protection scope of this application.

Claims (6)

1. An image class increment learning algorithm integrating uncertainty estimation and increment stage discrimination is characterized in that the method comprises the steps of performing multi-stage increment learning training on an image classification model to identify images of new classes and old classes, obtaining a classification head corresponding to each stage after each stage training, and training each stage by using an image data set of different classes, wherein each stage comprises the following steps:
1) A large-scale pre-training network (such as a CLIP network) is used as a feature extraction network of a classification model, parameters of the feature extraction network are frozen, data of a base class is input into the feature extraction network, and a cross entropy loss function L is used CE Training classification head parameters of the basic class stage, wherein the number of classes of the classification head is the number N of classes of the basic class data 0
2) For training in the kth incremental stage, freezing parameters of the feature extraction network, reserving classification heads after training in the previous incremental stage, inputting new class data into the feature extraction network, and using a cross entropy loss function L' CE Training the classification head parameters of the increment stage, wherein the classification head class number is N of the class number of the increment data of the stage k
3) In the test stage, the classification uncertainty H of the image x to be classified on the classification head of the kth increment stage is calculated k (x) Incremental stage to which classification head with minimum classification uncertainty belongsAs an incremental stage to which the image belongs;
4) Will be the firstThe classification head of each increment stage has the highest classification confidence class as the classification result of image x +.>
2. The image class increment learning algorithm combining uncertainty estimation and increment stage discrimination according to claim 1, wherein the cross entropy loss function L in step 1) CE The calculation formula of (2) is as follows:
wherein y is i Representing the output of the model classification head g i Representing the true tag value, N 0 The number of classes that are the classification heads of the base class stage.
3. The image class delta learning algorithm incorporating uncertainty estimation and delta phase discrimination as set forth in claim 1, wherein the cross entropy loss function L 'in step 2)' CE The calculation formula of (2) is as follows:
wherein y is i Representing the output of the model classification head g i Representing the true tag value, N k The number of categories for the category header for the kth incremental stage.
4. The image class increment learning algorithm combining uncertainty estimation and increment stage discrimination according to claim 1, wherein in step 3) the classification uncertainty of image x on the kth classification headH k (x) The calculation formula of (2) is as follows:
wherein y is ki (x) An ith output on the classification head representing image x at the kth incremental stage, N k Representing the number of categories of the classification header for the kth incremental stage.
5. The image class increment learning algorithm combining uncertainty estimation and increment stage discrimination according to claim 1, wherein the increment stage to which the classification head with the smallest classification uncertainty belongs in step 3)The calculation formula of (2) is as follows:
wherein n represents the number of stages of the increment stage up to now, and when the value of the increment stage k is 0, the increment stage is a base class stage and is the starting point of the increment learning model training.
6. The image class increment learning algorithm combining uncertainty estimation and increment stage discrimination according to claim 1, wherein the classification result of the image x in step 4) is thatThe calculation formula of (2) is as follows:
wherein,representing image x at delta stage +.>I-th output on classification head, < ->Representing delta phase +.>The number of categories of the classification header.
CN202311034343.5A 2023-08-16 2023-08-16 Image class increment learning algorithm integrating uncertainty estimation and increment stage discrimination Pending CN117079024A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311034343.5A CN117079024A (en) 2023-08-16 2023-08-16 Image class increment learning algorithm integrating uncertainty estimation and increment stage discrimination

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311034343.5A CN117079024A (en) 2023-08-16 2023-08-16 Image class increment learning algorithm integrating uncertainty estimation and increment stage discrimination

Publications (1)

Publication Number Publication Date
CN117079024A true CN117079024A (en) 2023-11-17

Family

ID=88703594

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311034343.5A Pending CN117079024A (en) 2023-08-16 2023-08-16 Image class increment learning algorithm integrating uncertainty estimation and increment stage discrimination

Country Status (1)

Country Link
CN (1) CN117079024A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117746140A (en) * 2023-12-21 2024-03-22 哈尔滨工业大学 Small sample class increment image classification method based on prompt word fine adjustment and feature replay
CN118172599A (en) * 2024-03-14 2024-06-11 上海交通大学 Mobile terminal new category learning and lightweight updating migration method based on parameter freezing

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117746140A (en) * 2023-12-21 2024-03-22 哈尔滨工业大学 Small sample class increment image classification method based on prompt word fine adjustment and feature replay
CN117746140B (en) * 2023-12-21 2025-01-10 哈尔滨工业大学 Incremental image classification method based on small sample classes based on cue word fine-tuning and feature replay
CN118172599A (en) * 2024-03-14 2024-06-11 上海交通大学 Mobile terminal new category learning and lightweight updating migration method based on parameter freezing

Similar Documents

Publication Publication Date Title
CN110909673B (en) Pedestrian re-identification method based on natural language description
CN109977918B (en) An Optimization Method for Object Detection and Localization Based on Unsupervised Domain Adaptation
CN108171209B (en) Face age estimation method for metric learning based on convolutional neural network
CN117079024A (en) Image class increment learning algorithm integrating uncertainty estimation and increment stage discrimination
CN106778796A (en) Human motion recognition method and system based on hybrid cooperative model training
CN106951825A (en) A kind of quality of human face image assessment system and implementation method
CN110163117B (en) Pedestrian re-identification method based on self-excitation discriminant feature learning
CN113963165B (en) Small sample image classification method and system based on self-supervision learning
Zhang et al. Self-guided adaptation: Progressive representation alignment for domain adaptive object detection
CN113743505A (en) An improved SSD object detection method based on self-attention and feature fusion
CN110097060A (en) A kind of opener recognition methods towards trunk image
CN109753897A (en) Behavior recognition method based on memory unit reinforcement-temporal dynamic learning
CN114529752A (en) Sample increment learning method based on deep neural network
CN117407772A (en) Method and system for classifying training multi-element time sequence data by supervising and comparing learning network model
CN116543269B (en) Cross-domain small-sample fine-grained image recognition method and its model based on self-supervision
CN111783688A (en) A classification method of remote sensing image scene based on convolutional neural network
CN116630718A (en) A Prototype-Based Low Perturbation Image-like Incremental Learning Algorithm
CN114170426A (en) Algorithm model for classifying rare tumor category small samples based on cost sensitivity
CN117726887A (en) Method, device and equipment for processing target domain data based on context awareness
Kuppusamy et al. Enriching the multi-object detection using convolutional neural network in macro-image
CN114492634B (en) Fine granularity equipment picture classification and identification method and system
CN106447691A (en) Weighted extreme learning machine video target tracking method based on weighted multi-example learning
Sarikabuta et al. Impacts of layer sizes in deep residual-learning convolutional neural network on flower image classification with different class sizes
CN118628813A (en) Passive domain adaptive image recognition method based on transferable semantic knowledge
CN108846349A (en) A kind of face identification method based on dynamic Spiking neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication