[go: up one dir, main page]

CN113177486B - Identification method of Odonata insects based on region proposal network - Google Patents

Identification method of Odonata insects based on region proposal network Download PDF

Info

Publication number
CN113177486B
CN113177486B CN202110480792.7A CN202110480792A CN113177486B CN 113177486 B CN113177486 B CN 113177486B CN 202110480792 A CN202110480792 A CN 202110480792A CN 113177486 B CN113177486 B CN 113177486B
Authority
CN
China
Prior art keywords
odonata
network model
images
deep convolutional
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110480792.7A
Other languages
Chinese (zh)
Other versions
CN113177486A (en
Inventor
皮家甜
于昕
彭明杰
吴志友
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing Normal University
Original Assignee
Chongqing Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing Normal University filed Critical Chongqing Normal University
Priority to CN202110480792.7A priority Critical patent/CN113177486B/en
Publication of CN113177486A publication Critical patent/CN113177486A/en
Application granted granted Critical
Publication of CN113177486B publication Critical patent/CN113177486B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a dragonfly order insect identification method based on a regional suggested network, which comprises the following steps: s1, cleaning and sorting images of insects in the order of dragonfly to obtain a data set of the images of the insects in the order of dragonfly; s2, enhancing the data set of the dragonfly order insect image to obtain an enhanced data set; s3, dividing the enhanced data set to obtain a training set, a verification set and a test set of the dragonfly order insect images; s4, constructing a deep convolutional network model based on the regional suggestion network; s5, training a deep convolution network model by using the training set and the verification set to obtain a trained network model; and S6, inputting the test set into the trained network model, and outputting to obtain a classification result of the test set. The method can enable the identification processing to be simple and rapid, save a large amount of labor cost, and solve the problem of difficult identification caused by complex background of the dragonfly-order insect pictures in natural environment.

Description

基于区域建议网络的蜻蜓目昆虫识别方法Recognition method of Odonata insects based on region proposal network

技术领域technical field

本发明涉及识别领域,具体涉及一种基于区域建议网络的蜻蜓目昆虫识别方法。The invention relates to the field of identification, in particular to a method for identifying Odonata insects based on a region suggestion network.

背景技术Background technique

现存的蜻蜓目自动识别算法以人为手工设计的特征作为分类依据,使用传统的识别方式构建识别框架,仅仅能够识别数种蜻蜓的标本图片,且识别率较低,对处于自然环境下背景复杂的蜻蜓图片不具备识别能力;The existing Odonata automatic identification algorithm uses the features designed by humans as the classification basis, and uses the traditional identification method to construct the identification frame, which can only identify the specimen pictures of several species of dragonflies, and the identification rate is low, which is very difficult for those with complex backgrounds in the natural environment. Dragonfly pictures do not have the ability to identify;

现存的昆虫自动识别算法往往分为两个步骤来实现。第一步,检测。即先针对待识别目标进行检测算法的实现,这个过程依赖大量的手工标注信息,耗时耗力,使得前期成本很高。第二步,识别。分为两种方式,其一,针对第一步中的检测算法的检测结果制作数据集,以此训练相应的分类器进行分类,这种方式的效果会受限于检测算法的性能;其二,对数据集进行手工框选分割,在保证训练图像包含完整的待识别目标的前提下训练分类器,在测试时再使用检测算法分割图像,而这种方式同样耗时耗力,增加了算法的人工成本。Existing insect automatic identification algorithms are often implemented in two steps. The first step is detection. That is, the detection algorithm is firstly implemented for the target to be identified. This process relies on a large amount of manual annotation information, which is time-consuming and labor-intensive, making the initial cost very high. The second step is to identify. There are two ways. One is to create a data set based on the detection results of the detection algorithm in the first step, so as to train the corresponding classifier for classification. The effect of this method will be limited by the performance of the detection algorithm. , perform manual box selection and segmentation on the data set, train the classifier on the premise that the training image contains the complete target to be recognized, and then use the detection algorithm to segment the image during testing, which is also time-consuming and labor-intensive, increasing the algorithm labor costs.

发明内容SUMMARY OF THE INVENTION

有鉴于此,本发明的目的是克服现有技术中的缺陷,提供基于区域建议网络的蜻蜓目昆虫识别方法,能够使得识别处理简单快捷,节省了大量的人力成本,解决了自然环境下蜻蜓目昆虫图片背景复杂导致识别困难的问题。In view of this, the purpose of the present invention is to overcome the defects in the prior art, and provide a method for identifying Odonata insects based on a regional suggestion network, which can make the identification process simple and fast, save a lot of labor costs, and solve the problem of Odonata in the natural environment. The complex background of insect pictures makes it difficult to identify.

本发明的基于区域建议网络的蜻蜓目昆虫识别方法,包括如下步骤:The Odonata insect identification method based on the regional suggestion network of the present invention comprises the following steps:

S1.采集蜻蜓目昆虫图像,并对蜻蜓目昆虫图像进行清洗和整理,得到蜻蜓目昆虫图像的数据集;S1. Collect images of Odonata insects, clean and organize the images of Odonata insects, and obtain a dataset of Odonata insect images;

S2.对蜻蜓目昆虫图像的数据集进行增强处理,得到增强后的数据集;S2. Enhance the dataset of dragonfly images to obtain an enhanced dataset;

S3.对所述增强后的数据集按照设定比例进行划分,得到蜻蜓目昆虫图像的训练集、验证集以及测试集;S3. Divide the enhanced data set according to a set ratio to obtain a training set, a verification set and a test set of the Odonata images;

S4.构建基于区域建议网络的深度卷积网络模型;S4. Build a deep convolutional network model based on the region proposal network;

S5.设置训练参数,并使用所述训练集以及验证集对深度卷积网络模型进行训练,得到训练后的网络模型;S5. set training parameters, and use the training set and the verification set to train the deep convolutional network model to obtain the trained network model;

S6.将所述测试集输入到训练后的网络模型,输出得到所述测试集的分类结果。S6. Input the test set into the trained network model, and output the classification result of the test set.

进一步,所述蜻蜓目昆虫图像为自然环境下的蜻蜓目昆虫图像。Further, the Odonata insect image is an Odonata insect image in a natural environment.

进一步,步骤S2中,对蜻蜓目昆虫图像的数据集进行增强处理,具体包括:对所述数据集进行随机水平翻转以及对所述数据集进行随机中心裁剪。Further, in step S2, the enhancement processing is performed on the data set of Odonata images, which specifically includes: random horizontal flipping of the data set and random center cropping of the data set.

进一步,所述步骤S4,具体包括:Further, the step S4 specifically includes:

S41.构建区域建议网络;S41. Build a regional proposal network;

S42.将ResNet50作为深度卷积网络模型的特征提取网络,并将区域建议网络作为深度卷积网络模型的特征筛选网络;S42. Use ResNet50 as the feature extraction network of the deep convolutional network model, and use the region proposal network as the feature screening network of the deep convolutional network model;

S43.确定深度卷积网络模型的损失函数。S43. Determine the loss function of the deep convolutional network model.

进一步,根据如下公式确定深度卷积网络模型的损失函数L:Further, the loss function L of the deep convolutional network model is determined according to the following formula:

L=L1+μL2L=L 1 +μL 2 ;

其中,L1为交叉熵损失函数;L2为改进的Focal Loss;μ为设定系数;Among them, L 1 is the cross entropy loss function; L 2 is the improved Focal Loss; μ is the setting coefficient;

所述

Figure BDA0003048485850000021
yi为当前样本的真实结果,
Figure BDA0003048485850000022
为所选子区域的预测结果;said
Figure BDA0003048485850000021
y i is the real result of the current sample,
Figure BDA0003048485850000022
is the prediction result of the selected sub-region;

所述

Figure BDA0003048485850000023
Figure BDA0003048485850000024
为当前样本的预测概率,α与γ分别为控制参数;said
Figure BDA0003048485850000023
Figure BDA0003048485850000024
is the predicted probability of the current sample, α and γ are the control parameters;

所述

Figure BDA0003048485850000025
id为样本编号,classes为样本的总标签数。said
Figure BDA0003048485850000025
id is the sample number, and classes is the total number of labels of the sample.

进一步,步骤S5中,设置训练参数,并使用所述训练集以及验证集对深度卷积网络模型进行训练,具体包括:Further, in step S5, set training parameters, and use the training set and the verification set to train the deep convolutional network model, specifically including:

S51.将在ImageNet数据集上的预训练权重模型作为ResNet50的初始化权重模型;S51. Use the pre-trained weight model on the ImageNet dataset as the initialization weight model of ResNet50;

S52.设置训练集与验证集中蜻蜓目昆虫图像的输入大小,并设置子区域的预设大小;S52. Set the input size of the Odonata images in the training set and the validation set, and set the preset size of the sub-region;

S53.使用Batch Normalization来实现正则化,并采用随机梯度下降算法对深度卷积网络模型进行优化。S53. Use Batch Normalization to achieve regularization, and use stochastic gradient descent to optimize the deep convolutional network model.

本发明的有益效果是:本发明公开的一种基于区域建议网络的蜻蜓目昆虫识别方法,通过以ResNet50为特征提取网络,无需进行手工特征的设计和提取,并且设计卷积神经网络作为区域建议网络来进行特征筛选,增强了模型从复杂背景中提取有效特征的能力,解决了自然环境下蜻蜓目昆虫图片背景复杂导致识别困难的问题。The beneficial effects of the present invention are as follows: a method for identifying Odonata insects based on a region suggestion network disclosed in the present invention, by using ResNet50 as a feature extraction network, no manual feature design and extraction are required, and a convolutional neural network is designed as a region suggestion The network is used for feature screening, which enhances the model's ability to extract effective features from complex backgrounds, and solves the problem of difficult identification caused by complex backgrounds in Odonata pictures in natural environments.

附图说明Description of drawings

下面结合附图和实施例对本发明作进一步描述:Below in conjunction with accompanying drawing and embodiment, the present invention is further described:

图1为本发明的方法流程示意图;Fig. 1 is the method flow schematic diagram of the present invention;

图2为本发明的所述方法整体网络框架图;2 is an overall network frame diagram of the method of the present invention;

图3为本发明的区域建议网络示意图;FIG. 3 is a schematic diagram of an area proposal network of the present invention;

图4为本发明的深度可分离卷积示意图;4 is a schematic diagram of a depthwise separable convolution of the present invention;

图5为本发明的池化与反池化示意图。FIG. 5 is a schematic diagram of pooling and de-pooling of the present invention.

具体实施方式Detailed ways

以下结合说明书附图对本发明做出进一步的说明,如图1所示:The present invention is further described below in conjunction with the accompanying drawings, as shown in Figure 1:

本发明的基于区域建议网络的蜻蜓目昆虫识别方法,包括如下步骤:The Odonata insect identification method based on the regional suggestion network of the present invention comprises the following steps:

S1.采集蜻蜓目昆虫图像,并对蜻蜓目昆虫图像进行清洗和整理,得到蜻蜓目昆虫图像的数据集;其中,蜻蜓目昆虫图像数据的获取方式包括野外拍摄、实验室拍摄和网络下载等多种方式,因此最后得到的蜻蜓目昆虫图像的数据集不具有给定条件下的各类蜻蜓目昆虫图像数量均等的特点,整体数据呈长尾分布。S1. Collect images of Odonata insects, clean and organize the images of Odonata insects, and obtain a dataset of Odonata insect images; among which, the acquisition methods of Odonata insect image data include field shooting, laboratory shooting, and network download, etc. Therefore, the final data set of Odonata images does not have the characteristic that the number of images of various Odonata insects is equal under the given conditions, and the overall data has a long-tailed distribution.

S2.对蜻蜓目昆虫图像的数据集进行增强处理,得到增强后的数据集;S2. Enhance the dataset of dragonfly images to obtain an enhanced dataset;

S3.对所述增强后的数据集按照设定比例进行划分,得到蜻蜓目昆虫图像的训练集、验证集以及测试集;其中,所述设定比例为4.5:4.5:1。S3. Divide the enhanced data set according to a set ratio to obtain a training set, a verification set and a test set of Odonata images; wherein, the set ratio is 4.5:4.5:1.

S4.构建基于区域建议网络的深度卷积网络模型;其中,通过构建基于区域建议网络的深度卷积网络模型,省去了先对蜻蜓目昆虫进行检测的步骤,从而避免了对数据进行大量的人工标注工作,也使的所述深度卷积网络模型能够直接识别待测图像中的蜻蜓种类;S4. Construct a deep convolutional network model based on the region proposal network; wherein, by constructing a deep convolutional network model based on the region proposal network, the step of firstly detecting Odonata insects is omitted, thereby avoiding a large amount of data processing. The manual annotation work also enables the deep convolutional network model to directly identify the dragonfly species in the image to be tested;

S5.设置训练参数,并使用所述训练集以及验证集对深度卷积网络模型进行训练,得到训练后的网络模型;S5. set training parameters, and use the training set and the verification set to train the deep convolutional network model to obtain the trained network model;

S6.将所述测试集输入到训练后的网络模型,输出得到所述测试集的分类结果。S6. Input the test set into the trained network model, and output the classification result of the test set.

本实施例中,所述蜻蜓目昆虫图像为自然环境下的蜻蜓目昆虫图像。其中,采集的蜻蜓目昆虫图像中个别种类包含极少量的实验室标本图像。In this embodiment, the Odonata insect image is an Odonata insect image in a natural environment. Among them, individual species in the collected Odonata images contain very few laboratory specimen images.

本实施例中,步骤S2中,对蜻蜓目昆虫图像的数据集进行增强处理,具体包括:对所述数据集进行随机水平翻转以及对所述数据集进行随机中心裁剪。In this embodiment, in step S2, the enhancement processing is performed on the data set of Odonata images, which specifically includes: random horizontal flipping of the data set and random center cropping of the data set.

本实施例中,所述步骤S4,具体包括:In this embodiment, the step S4 specifically includes:

S41.构建区域建议网络;其中,所述区域建议网络简称为RPN(region proposalnetwork),如图3所示,黄色虚边块代表平均池化,黄色无边块代表最大反池化,绿色无边块代表深度可分离卷积;对于大小为H×W×C的输入图片,在通过所述RPN后,得到一组预先设定好数量(k)和大小(h×w)的h×w×C×k个子区域,将这k个区域与原图像进行相加来作为最终分类网络的输入。S41. Build a region proposal network; wherein, the region proposal network is abbreviated as RPN (region proposal network). As shown in Figure 3, the yellow imaginary blocks represent average pooling, the yellow borderless blocks represent maximum de-pooling, and the green borderless blocks represent Depthwise separable convolution; for an input image of size H×W×C, after passing through the RPN, a set of h×w×C× with a preset number (k) and size (h×w) is obtained k sub-regions, which are added to the original image as the input of the final classification network.

S42.将ResNet50作为深度卷积网络模型的特征提取网络,并将区域建议网络作为深度卷积网络模型的特征筛选网络;具体地,如图4、5所示,对输入图片F∈RH×W×C进行M个不同预设尺度的子区域选取,通过3×3的深度卷积加1×1的点卷积,次接3×3的平均池化,再接3×3的最大反池化(从此处单分一支1×1的卷积),随后又是一组深度可分离卷积(此处单分一支1×1的卷积),后接3×3标准卷积,紧接1×1卷积,采样过程使用ReLu函数为激活函数,最后与两支路级联得到特征图FR,然后进行预分类。再根据预分类结果,从M个区域中选取置信度最高的k个区域作为RPN筛选的子区域作为输出S={R1,R2,…,Rk}。S42. Use ResNet50 as the feature extraction network of the deep convolutional network model, and use the region proposal network as the feature screening network of the deep convolutional network model; specifically, as shown in Figures 4 and 5, for the input picture F∈R H× W×C selects M sub-regions of different preset scales, through 3×3 depth convolution plus 1×1 point convolution, followed by 3×3 average pooling, and then 3×3 maximum inversion Pooling (a single 1×1 convolution from here), followed by a set of depthwise separable convolutions (a single 1×1 convolution here), followed by a 3×3 standard convolution , followed by 1×1 convolution, the sampling process uses the ReLu function as the activation function, and finally cascades with two branches to obtain the feature map F R , and then pre-classifies. Then, according to the pre-classification results, the k regions with the highest confidence are selected from the M regions as the sub-regions screened by the RPN as the output S={R 1 , R 2 ,...,R k }.

S43.确定深度卷积网络模型的损失函数。其中,通过引入所述损失函数,并加入超参数来控制整体网络损失中的占比,解决了所述数据集存在各类样本数量不均衡的问题。具体地,对输入图片F∈RH×W×C和输出S={R1,R2,…,Rk},在使用ResNet50进行特征提取后,将得到的特征图相加后进行分类,所述分类损失采用所述损失函数进行处理。S43. Determine the loss function of the deep convolutional network model. Among them, by introducing the loss function and adding hyperparameters to control the proportion of the overall network loss, the problem of uneven number of samples of various types in the data set is solved. Specifically, for the input picture F∈R H×W×C and the output S={R 1 ,R 2 ,...,R k }, after using ResNet50 for feature extraction, the obtained feature maps are added for classification, The classification loss is processed using the loss function.

本实施例中,根据如下公式确定深度卷积网络模型的损失函数L:In this embodiment, the loss function L of the deep convolutional network model is determined according to the following formula:

L=L1+μL2L=L 1 +μL 2 ;

其中,L1为交叉熵损失函数;L2为改进的Focal Loss;μ为设定系数;Among them, L 1 is the cross entropy loss function; L 2 is the improved Focal Loss; μ is the setting coefficient;

所述

Figure BDA0003048485850000051
yi为当前样本的真实结果,
Figure BDA0003048485850000052
为所选子区域的预测结果;said
Figure BDA0003048485850000051
y i is the real result of the current sample,
Figure BDA0003048485850000052
is the prediction result of the selected sub-region;

所述

Figure BDA0003048485850000053
Figure BDA0003048485850000054
为当前样本的预测概率,α与γ分别为控制参数,所述参数用来调节所述数据集中易分类样本的损失计算权重,本实施例中,设置参数α为0.5,参数γ为2;said
Figure BDA0003048485850000053
Figure BDA0003048485850000054
is the prediction probability of the current sample, α and γ are control parameters, respectively, and the parameters are used to adjust the loss calculation weight of the easy-to-classify samples in the data set. In this embodiment, the parameter α is set to 0.5, and the parameter γ is set to 2;

所述

Figure BDA0003048485850000055
id为样本编号,classes为样本的总标签数。在本实施例中,在制作所述数据集时,按各类所含样本数的多少降序来设置样本编号(id),即第一类数量最多,最后一类数量最少。当当前训练样本的标签在样本的总标签数(classes)的前一半时,μ=0.5,反之μ=1。said
Figure BDA0003048485850000055
id is the sample number, and classes is the total number of labels of the sample. In this embodiment, when creating the data set, the sample numbers (id) are set in descending order of the number of samples contained in each category, that is, the first category has the largest number and the last category has the smallest number. When the label of the current training sample is in the first half of the total number of labels (classes) of the sample, μ=0.5, otherwise μ=1.

本实施例中,步骤S5中,设置训练参数,并使用所述训练集以及验证集对深度卷积网络模型进行训练,具体包括:In this embodiment, in step S5, training parameters are set, and the training set and the verification set are used to train the deep convolutional network model, which specifically includes:

S51.将在ImageNet数据集上的预训练权重模型作为ResNet50的初始化权重模型;S51. Use the pre-trained weight model on the ImageNet dataset as the initialization weight model of ResNet50;

S52.设置训练集与验证集中蜻蜓目昆虫图像的输入大小,并设置子区域的预设大小;S52. Set the input size of the Odonata images in the training set and the validation set, and set the preset size of the sub-region;

S53.使用Batch Normalization来实现正则化,并采用随机梯度下降算法对深度卷积网络模型进行优化。S53. Use Batch Normalization to achieve regularization, and use stochastic gradient descent to optimize the deep convolutional network model.

具体地,在训练过程中,图像的输入大小(Input size)设置为448×448(像素值),所述子区域预设大小为48×48,96×96,192×192(像素值),k的取值为3,使用BN(BatchNormalization)来实现正则化。初始学习率为0.001,优化器为Momentum SGD。NMS阈值为0.25,权重衰减为0.0001,迭代epoch为200。Specifically, in the training process, the input size of the image (Input size) is set to 448×448 (pixel value), and the preset sizes of the sub-regions are 48×48, 96×96, 192×192 (pixel value), The value of k is 3, and BN (BatchNormalization) is used to achieve regularization. The initial learning rate is 0.001 and the optimizer is Momentum SGD. The NMS threshold is 0.25, the weight decay is 0.0001, and the iteration epoch is 200.

最后说明的是,以上实施例仅用以说明本发明的技术方案而非限制,尽管参照较佳实施例对本发明进行了详细说明,本领域的普通技术人员应当理解,可以对本发明的技术方案进行修改或者等同替换,而不脱离本发明技术方案的宗旨和范围,其均应涵盖在本发明的权利要求范围当中。Finally, it should be noted that the above embodiments are only used to illustrate the technical solutions of the present invention and not to limit them. Although the present invention has been described in detail with reference to the preferred embodiments, those of ordinary skill in the art should understand that the technical solutions of the present invention can be Modifications or equivalent substitutions without departing from the spirit and scope of the technical solutions of the present invention should be included in the scope of the claims of the present invention.

Claims (4)

1.一种基于区域建议网络的蜻蜓目昆虫识别方法,其特征在于:包括如下步骤:1. an Odonata insect identification method based on regional suggestion network, is characterized in that: comprise the steps: S1.采集蜻蜓目昆虫图像,并对蜻蜓目昆虫图像进行清洗和整理,得到蜻蜓目昆虫图像的数据集;S1. Collect images of Odonata insects, clean and organize the images of Odonata insects, and obtain a dataset of Odonata insect images; S2.对蜻蜓目昆虫图像的数据集进行增强处理,得到增强后的数据集;S2. Enhance the dataset of dragonfly images to obtain an enhanced dataset; S3.对所述增强后的数据集按照设定比例进行划分,得到蜻蜓目昆虫图像的训练集、验证集以及测试集;S3. Divide the enhanced data set according to a set ratio to obtain a training set, a verification set and a test set of the Odonata images; S4.构建基于区域建议网络的深度卷积网络模型;S4. Build a deep convolutional network model based on the region proposal network; 所述步骤S4,具体包括:The step S4 specifically includes: S41.构建区域建议网络;S41. Build a regional proposal network; S42.将ResNet50作为深度卷积网络模型的特征提取网络,并将区域建议网络作为深度卷积网络模型的特征筛选网络;具体包括:对输入图片进行M个不同预设尺度的子区域选取,通过深度卷积加点卷积,次接平均池化,再接最大反池化,从此处单分一支卷积,随后又是一组深度可分离卷积,此处单分一支卷积,后接标准卷积,紧接卷积,采样过程使用ReLu函数为激活函数,最后与两支路级联得到特征图,然后进行预分类,再根据预分类结果,从M个区域中选取置信度最高的k个区域作为RPN筛选的子区域,输出S={R1,R2,…,Rk};S42. Use ResNet50 as the feature extraction network of the deep convolutional network model, and use the region proposal network as the feature screening network of the deep convolutional network model; specifically including: selecting M sub-regions of different preset scales for the input picture, Depth convolution plus point convolution, followed by average pooling, followed by maximum unpooling, from here a single convolution, followed by a group of depthwise separable convolutions, here a single convolution, and then a single convolution. The standard convolution is followed by the convolution. The sampling process uses the ReLu function as the activation function. Finally, it is cascaded with two branches to obtain a feature map, and then pre-classification is performed. Then, according to the pre-classification results, the M areas with the highest confidence are selected. The k regions of are used as sub-regions of RPN screening, and output S={R 1 , R 2 ,...,R k }; S43.确定深度卷积网络模型的损失函数;S43. Determine the loss function of the deep convolutional network model; 根据如下公式确定深度卷积网络模型的损失函数L:The loss function L of the deep convolutional network model is determined according to the following formula: L=L1+μL2L=L 1 +μL 2 ; 其中,L1为交叉熵损失函数;L2为改进的Focal Loss;μ为设定系数;Among them, L 1 is the cross entropy loss function; L 2 is the improved Focal Loss; μ is the setting coefficient; 所述
Figure FDA0003588006480000011
yi为当前样本的真实结果,
Figure FDA0003588006480000012
为所选子区域的预测结果;
said
Figure FDA0003588006480000011
y i is the real result of the current sample,
Figure FDA0003588006480000012
is the prediction result of the selected sub-region;
所述
Figure FDA0003588006480000013
Figure FDA0003588006480000014
为当前样本的预测概率,α与γ分别为控制参数;
said
Figure FDA0003588006480000013
Figure FDA0003588006480000014
is the predicted probability of the current sample, α and γ are the control parameters;
所述
Figure FDA0003588006480000021
id为样本编号,按各类所含样本数的多少降序来设置样本编号,classes为样本的总标签数;
said
Figure FDA0003588006480000021
id is the sample number, set the sample number in descending order of the number of samples contained in each category, and classes is the total number of labels of the sample;
S5.设置训练参数,并使用所述训练集以及验证集对深度卷积网络模型进行训练,得到训练后的网络模型;S5. set training parameters, and use the training set and the verification set to train the deep convolutional network model to obtain the trained network model; S6.将所述测试集输入到训练后的网络模型,输出得到所述测试集的分类结果。S6. Input the test set into the trained network model, and output the classification result of the test set.
2.根据权利要求1所述的基于区域建议网络的蜻蜓目昆虫识别方法,其特征在于:所述蜻蜓目昆虫图像为自然环境下的蜻蜓目昆虫图像。2 . The method for identifying Odonata insects based on a region proposal network according to claim 1 , wherein the Odonata insect images are images of Odonata insects in a natural environment. 3 . 3.根据权利要求1所述的基于区域建议网络的蜻蜓目昆虫识别方法,其特征在于:步骤S2中,对蜻蜓目昆虫图像的数据集进行增强处理,具体包括:对所述数据集进行随机水平翻转以及对所述数据集进行随机中心裁剪。3. The method for identifying Odonata insects based on a region suggestion network according to claim 1, wherein in step S2, the data set of the Odonata insect images is enhanced, specifically comprising: randomly performing a randomization process on the data set. Horizontal flipping and random center cropping of the dataset. 4.根据权利要求1所述的基于区域建议网络的蜻蜓目昆虫识别方法,其特征在于:步骤S5中,设置训练参数,并使用所述训练集以及验证集对深度卷积网络模型进行训练,具体包括:4. the Odonata insect identification method based on the regional suggestion network according to claim 1, is characterized in that: in step S5, set training parameter, and use described training set and verification set to train deep convolutional network model, Specifically include: S51.将在ImageNet数据集上的预训练权重模型作为ResNet50的初始化权重模型;S51. Use the pre-trained weight model on the ImageNet dataset as the initialization weight model of ResNet50; S52.设置训练集与验证集中蜻蜓目昆虫图像的输入大小,并设置子区域的预设大小;S52. Set the input size of the Odonata images in the training set and the validation set, and set the preset size of the sub-region; S53.使用Batch Normalization来实现正则化,并采用随机梯度下降算法对深度卷积网络模型进行优化。S53. Use Batch Normalization to achieve regularization, and use stochastic gradient descent algorithm to optimize the deep convolutional network model.
CN202110480792.7A 2021-04-30 2021-04-30 Identification method of Odonata insects based on region proposal network Active CN113177486B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110480792.7A CN113177486B (en) 2021-04-30 2021-04-30 Identification method of Odonata insects based on region proposal network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110480792.7A CN113177486B (en) 2021-04-30 2021-04-30 Identification method of Odonata insects based on region proposal network

Publications (2)

Publication Number Publication Date
CN113177486A CN113177486A (en) 2021-07-27
CN113177486B true CN113177486B (en) 2022-06-03

Family

ID=76925719

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110480792.7A Active CN113177486B (en) 2021-04-30 2021-04-30 Identification method of Odonata insects based on region proposal network

Country Status (1)

Country Link
CN (1) CN113177486B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114170530A (en) * 2021-11-11 2022-03-11 国网福建省电力有限公司漳州供电公司 Auxiliary acquisition method and system of UAV line patrol image based on resolution reconstruction

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111126385A (en) * 2019-12-13 2020-05-08 哈尔滨工程大学 Deep learning intelligent identification method for deformable living body small target
CN111652247A (en) * 2020-05-28 2020-09-11 大连海事大学 A Dipteran Insect Recognition Method Based on Deep Convolutional Neural Networks
CN111898406A (en) * 2020-06-05 2020-11-06 东南大学 Face detection method based on focal loss and multi-task cascade
CN111931581A (en) * 2020-07-10 2020-11-13 威海精讯畅通电子科技有限公司 Agricultural pest identification method based on convolutional neural network, terminal and readable storage medium
CN112070043A (en) * 2020-09-15 2020-12-11 常熟理工学院 Safety helmet wearing convolutional network based on feature fusion, training and detecting method
CN112116603A (en) * 2020-09-14 2020-12-22 中国科学院大学宁波华美医院 Pulmonary nodule false positive screening method based on multitask learning
WO2021008233A1 (en) * 2019-07-17 2021-01-21 上海商汤智能科技有限公司 Robot image enhancement method and apparatus, processor, device, medium and program
CN112288795A (en) * 2020-10-29 2021-01-29 深圳大学 Insect density calculation method and device based on fast-RCNN
CN112598657A (en) * 2020-12-28 2021-04-02 锋睿领创(珠海)科技有限公司 Defect detection method and device, model construction method and computer equipment

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10579897B2 (en) * 2017-10-02 2020-03-03 Xnor.ai Inc. Image based object detection
CN109359681B (en) * 2018-10-11 2022-02-11 西京学院 A method for identification of field crop diseases and insect pests based on improved fully convolutional neural network
CN111444952B (en) * 2020-03-24 2024-02-20 腾讯科技(深圳)有限公司 Sample recognition model generation method, device, computer equipment and storage medium
CN111476302B (en) * 2020-04-08 2023-03-24 北京工商大学 fast-RCNN target object detection method based on deep reinforcement learning
CN112464971A (en) * 2020-04-09 2021-03-09 丰疆智能软件科技(南京)有限公司 Method for constructing pest detection model
CN112257569B (en) * 2020-10-21 2021-11-19 青海城市云大数据技术有限公司 Target detection and identification method based on real-time video stream
CN112465819B (en) * 2020-12-18 2024-06-18 平安科技(深圳)有限公司 Image abnormal region detection method and device, electronic equipment and storage medium

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021008233A1 (en) * 2019-07-17 2021-01-21 上海商汤智能科技有限公司 Robot image enhancement method and apparatus, processor, device, medium and program
CN111126385A (en) * 2019-12-13 2020-05-08 哈尔滨工程大学 Deep learning intelligent identification method for deformable living body small target
CN111652247A (en) * 2020-05-28 2020-09-11 大连海事大学 A Dipteran Insect Recognition Method Based on Deep Convolutional Neural Networks
CN111898406A (en) * 2020-06-05 2020-11-06 东南大学 Face detection method based on focal loss and multi-task cascade
CN111931581A (en) * 2020-07-10 2020-11-13 威海精讯畅通电子科技有限公司 Agricultural pest identification method based on convolutional neural network, terminal and readable storage medium
CN112116603A (en) * 2020-09-14 2020-12-22 中国科学院大学宁波华美医院 Pulmonary nodule false positive screening method based on multitask learning
CN112070043A (en) * 2020-09-15 2020-12-11 常熟理工学院 Safety helmet wearing convolutional network based on feature fusion, training and detecting method
CN112288795A (en) * 2020-10-29 2021-01-29 深圳大学 Insect density calculation method and device based on fast-RCNN
CN112598657A (en) * 2020-12-28 2021-04-02 锋睿领创(珠海)科技有限公司 Defect detection method and device, model construction method and computer equipment

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
"Destruction and construction learning for fine-grained image recognition";CHEN Y等;《Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition》;20191231;第5157-5166页 *
"Learning to navigate for fine-grained classification";Yang Z等;《Proceedings of the 2018 European Conference on Computer Vision》;20181231;第438-454页 *
"基于RPN与B-CNN的细粒度图像分类算法研究";赵浩如等;《计算机应用与软件》;20190331;第36卷(第3期);第210-213页及264页 *
"基于改进残差网络的道口车辆分类方法";李宇昕等;<《激光与光电子学进展》;20210228;第58卷(第4期);第1-7页 *
"深度区域网络方法的细粒度图像分类";翁雨辰等;《中国图象图形学报》;20171231;第22卷(第11期);第1521-1531页 *

Also Published As

Publication number Publication date
CN113177486A (en) 2021-07-27

Similar Documents

Publication Publication Date Title
CN111310862B (en) Image enhancement-based deep neural network license plate positioning method in complex environment
WO2021203505A1 (en) Method for constructing pest detection model
CN106778835B (en) Remote sensing image airport target identification method fusing scene information and depth features
CN111814704B (en) Full convolution examination room target detection method based on cascade attention and point supervision mechanism
CN114092389A (en) A surface defect detection method for glass panels based on small sample learning
CN107133616A (en) A kind of non-division character locating and recognition methods based on deep learning
CN109800631A (en) Fluorescence-encoded micro-beads image detecting method based on masked areas convolutional neural networks
CN108846835A (en) The image change detection method of convolutional network is separated based on depth
CN109344851B (en) Image classification display method and device, analysis instrument and storage medium
CN108629369A (en) A kind of Visible Urine Sediment Components automatic identifying method based on Trimmed SSD
CN112686862B (en) Pest identification and counting method, system, device and readable storage medium
CN105955708A (en) Sports video lens classification method based on deep convolutional neural networks
CN112819821A (en) Cell nucleus image detection method
CN111914902B (en) Traditional Chinese medicine identification and surface defect detection method based on deep neural network
CN116012709B (en) High-resolution remote sensing image building extraction method and system
CN112862849A (en) Image segmentation and full convolution neural network-based field rice ear counting method
CN113313678A (en) Automatic sperm morphology analysis method based on multi-scale feature fusion
CN114140663A (en) Multi-scale attention and learning network-based pest identification method and system
CN112991280B (en) Visual detection method, visual detection system and electronic equipment
CN113177486B (en) Identification method of Odonata insects based on region proposal network
CN115147646A (en) A detection method of small-target pests in garden engineering based on super-resolution reconstruction and data enhancement
CN112465821A (en) Multi-scale pest image detection method based on boundary key point perception
CN111414951B (en) Method and device for subdividing images
CN113344919B (en) Method and system for detecting ceramic thermal shock damage degree based on convolutional neural network
CN107203788B (en) Medium-level visual drug image identification method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant