CN112506797B - A performance testing method for medical image recognition system - Google Patents
A performance testing method for medical image recognition system Download PDFInfo
- Publication number
- CN112506797B CN112506797B CN202011525218.0A CN202011525218A CN112506797B CN 112506797 B CN112506797 B CN 112506797B CN 202011525218 A CN202011525218 A CN 202011525218A CN 112506797 B CN112506797 B CN 112506797B
- Authority
- CN
- China
- Prior art keywords
- image
- test
- loss
- performance
- style
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000012360 testing method Methods 0.000 title claims abstract description 120
- 238000000034 method Methods 0.000 claims abstract description 51
- 238000011156 evaluation Methods 0.000 claims abstract description 25
- 238000011056 performance test Methods 0.000 claims abstract description 10
- 238000005215 recombination Methods 0.000 claims abstract 2
- 230000006798 recombination Effects 0.000 claims abstract 2
- 230000008859 change Effects 0.000 claims description 21
- 230000003416 augmentation Effects 0.000 claims description 18
- 230000009466 transformation Effects 0.000 claims description 14
- 230000006870 function Effects 0.000 claims description 13
- 230000008521 reorganization Effects 0.000 claims description 10
- 230000003190 augmentative effect Effects 0.000 claims description 8
- 101100153586 Caenorhabditis elegans top-1 gene Proteins 0.000 claims description 6
- 101100153591 Cricetulus griseus TOP1 gene Proteins 0.000 claims description 6
- 101100370075 Mus musculus Top1 gene Proteins 0.000 claims description 6
- 230000035945 sensitivity Effects 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 4
- 238000000844 transformation Methods 0.000 claims description 4
- 230000008439 repair process Effects 0.000 claims description 3
- 230000011218 segmentation Effects 0.000 claims description 3
- 238000012549 training Methods 0.000 claims description 3
- 230000006978 adaptation Effects 0.000 claims description 2
- 238000011076 safety test Methods 0.000 claims description 2
- 229910008048 Si-S Inorganic materials 0.000 claims 1
- 229910006336 Si—S Inorganic materials 0.000 claims 1
- 230000003042 antagnostic effect Effects 0.000 claims 1
- 239000000284 extract Substances 0.000 claims 1
- 230000008004 immune attack Effects 0.000 claims 1
- 239000011159 matrix material Substances 0.000 claims 1
- 238000005516 engineering process Methods 0.000 description 5
- 238000004458 analytical method Methods 0.000 description 4
- 238000011161 development Methods 0.000 description 4
- 230000018109 developmental process Effects 0.000 description 4
- 239000002131 composite material Substances 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 239000003814 drug Substances 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 239000013589 supplement Substances 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 238000003759 clinical diagnosis Methods 0.000 description 1
- 238000013170 computed tomography imaging Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000002595 magnetic resonance imaging Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000029052 metamorphosis Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 239000000047 product Substances 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 238000010998 test method Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/36—Prevention of errors by analysis, debugging or testing of software
- G06F11/3668—Testing of software
- G06F11/3672—Test management
- G06F11/3684—Test management for test design, e.g. generating new test cases
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/36—Prevention of errors by analysis, debugging or testing of software
- G06F11/3668—Testing of software
- G06F11/3672—Test management
- G06F11/3688—Test management for test execution, e.g. scheduling of test suites
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Computer Hardware Design (AREA)
- Quality & Reliability (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域technical field
本发明属于医学影像识别系统性能分析技术领域,特别涉及一种针对医学影像识别系统的性能测试方法。The invention belongs to the technical field of performance analysis of a medical image recognition system, in particular to a performance testing method for a medical image recognition system.
背景技术Background technique
医学影像识别系统在临床医学诊断中发挥着重要的作用,它大大改变了临床诊断模式,促进了临床医学的发展。智能医学影像识别是基于人工智能技术,对X线片、计算机断层扫描、磁共振成像等常用医学影像学技术扫描图像和手术视频进行分析处理的过程,其发展方向主要包括智能影像诊断、影像三维重建与配准、智能手术视频解析等。目前,对该领域的研究已取得一定的进展,正在逐步走向临床应用。因此,对医学影像识别系统性能的评估和测试对未来临床医学的发展尤为重要。FERET首次给识别算法设置了性能基准,定义了一系列的评测标准,极大地推进了识别技术的发展,其制定的评测标准和评价协议一直影响至今,为以后人脸识别技术的发展带来了深远的影响。虽然一般的图像识别系统有一些测试方案,但是针对医学影像识别系统的测试方案还没有被提出。而且由于当时的识别技术尚不成熟,参与FERET评测的识别系统也多是大学实验室里的原型系统,识别效果并不是非常令人满意。Medical image recognition system plays an important role in clinical medical diagnosis. It has greatly changed the clinical diagnosis mode and promoted the development of clinical medicine. Intelligent medical image recognition is a process of analyzing and processing scanned images and surgical videos of commonly used medical imaging technologies such as X-ray, computed tomography, and magnetic resonance imaging based on artificial intelligence technology. Reconstruction and registration, intelligent surgical video analysis, etc. At present, the research in this field has made some progress and is gradually moving towards clinical application. Therefore, the evaluation and testing of the performance of the medical image recognition system is particularly important for the development of clinical medicine in the future. For the first time, FERET set a performance benchmark for the recognition algorithm and defined a series of evaluation standards, which greatly promoted the development of recognition technology. profound influence. Although there are some test schemes for general image recognition systems, no test schemes for medical image recognition systems have been proposed. Moreover, since the recognition technology at that time was not yet mature, the recognition systems involved in the FERET evaluation were mostly prototype systems in university laboratories, and the recognition effect was not very satisfactory.
近年来,对识别模型进行分析,并自动进行性能分析的测试方法需求逐渐增加。随着深度学习技术快速发展,医学影像识别系统的性能指标也得到快速提高,大大提高了识别效率,所以如何测试这些模型的性能亟待解决。针对医学影像识别系统在测试中,需要考虑到生成的测试影像的视觉真实性,因此提出了对抗样本生成网络和实体、背景重组方法,充分保证生成样本的真实性;且医学识别系统对可靠性和安全性的测试要求更高,因此提出多角度测试方案,并将对抗样本应用到医学影像中,达到对医学识别模型更好分析的效果。In recent years, there has been a growing demand for test methods that analyze recognition models and automate performance analysis. With the rapid development of deep learning technology, the performance indicators of medical image recognition systems have also been rapidly improved, greatly improving the recognition efficiency, so how to test the performance of these models needs to be solved urgently. For the medical image recognition system in the test, it is necessary to consider the visual authenticity of the generated test image, so the adversarial sample generation network and the entity and background reorganization method are proposed to fully ensure the authenticity of the generated samples; Therefore, a multi-angle test scheme is proposed, and the adversarial samples are applied to medical images to achieve better analysis of the medical recognition model.
发明内容SUMMARY OF THE INVENTION
本发明提供了一种针对医学影像识别系统的性能测试方法,以解决现有技术中的问题。The present invention provides a performance testing method for a medical image recognition system to solve the problems in the prior art.
为了实现上述目的,本发明采用以下技术方案:In order to achieve the above object, the present invention adopts the following technical solutions:
一种针对医学影像识别系统的性能测试方法,性能测试方法中包括了:多类别影像测试数据生成模块,所述多类别影像测试数据生成模块包括对抗样本生成网络和实体、背景重组方法;多角度测试模块,所述多角度测试模块包括性能测试、可靠性测试和安全性测试;决策评估模块,所述决策评估模块对所输入的测试结果进行分析,判断模型性能,并给出详细的测试报告;A performance testing method for a medical image recognition system, the performance testing method includes: a multi-category image test data generation module, wherein the multi-category image test data generation module includes an adversarial sample generation network and an entity and background reorganization method; Test module, the multi-angle test module includes performance test, reliability test and safety test; decision evaluation module, the decision evaluation module analyzes the input test results, judges the model performance, and gives a detailed test report ;
网络输入一组待分类识别的图片,首次将图片输入到所述多类别影像测试数据生成模块中,经过图像增广后输入模型进行分类,并将分类结果输入至多角度测试模块,多角度测试模块对模型的学习结果进行测试,并将结果传递到决策评估模块,决策评估模块对所输入的测试结果进行分析,判断模型性能,并给出详细的测试报告。The network inputs a set of pictures to be classified and recognized, and the pictures are input into the multi-category image test data generation module for the first time, and after image augmentation, the model is input for classification, and the classification results are input into the multi-angle test module, which is a multi-angle test module. The learning results of the model are tested, and the results are transmitted to the decision evaluation module. The decision evaluation module analyzes the input test results, judges the performance of the model, and gives a detailed test report.
进一步的,所述对抗样本生成网络和实体、背景重组方法中包括对抗增广使用多损失混合对抗伪装,Further, the adversarial sample generation network and the entity and background reorganization method include adversarial augmentation using multi-loss mixed adversarial camouflage,
多损失函数表示为:Multiple loss functions Expressed as:
其中:λ表示对抗强度、表示对抗性损失、表示用于样式生成的样式损失、表示用于保存源图像内容的内容损失、表示用于确保扩展样本的平滑度的平滑度损失;Among them: λ represents the strength of confrontation, represents the adversarial loss, represents the style loss used for style generation, Indicates the content loss used to preserve the content of the source image, represents the smoothness loss used to ensure smoothness of the expanded samples;
用户定义现有图像、目标攻击区域和预期目标样式,在所需的区域中生成所需的样式,在每个步骤中向生成的扩展样本中添加附加的物理适应训练;The user defines the existing image, the target attack area and the expected target style, generates the desired style in the desired area, adds additional physical fitness training to the generated extended samples at each step;
两个图像之间的样式距离通过两个图像在样式表示中的差异来定义,The style distance between two images is defined by the difference in the style representation of the two images,
其中:为特征距离,l是风格层特征,Sl是提取风格表示的风格层集,是风格样式的特征提取器,是从的一组样式层提取的深层特征的Gram矩阵,xs是风格参考图像,x′是生成的对抗样本;in: is the feature distance, l is the style layer feature, S l is the style layer set for extracting style representation, is a style-style feature extractor, From A set of Gram matrices of deep features extracted by the style layer, x s is the style reference image, and x′ is the generated adversarial example;
用于样式生成的样式损失所生成参考样式中的增强图像的内容与原始图像的内容非常不同;具体如下,style loss for style generation The content of the enhanced image in the generated reference style is very different from the content of the original image; as follows,
其中:是内容损失,t是内容层特征,ct是提取内容表示的内容层集,是内容层的特征提取器,x是原始图像,x′是生成的对抗样本;in: is the content loss, t is the content layer feature, c t is the content layer set from which the content representation is extracted, is the feature extractor of the content layer, x is the original image, and x′ is the generated adversarial example;
通过减少相邻像素之间的变化,提高增强图像的平滑度;对于增强图像,平滑度损失定义如下,The smoothness of the enhanced image is improved by reducing the variation between adjacent pixels; for the enhanced image, the smoothness loss is defined as follows,
其中:xi,j为对抗样本(i,j)坐标处的像素值,xi+1,j为原始图像(i+1,j)坐标处的像素值,xi,j+1为原始图像(i,j+1)坐标处的像素值:Among them: x i, j is the pixel value at the coordinate of the adversarial sample (i, j), x i+1, j is the pixel value at the coordinate of the original image (i+1, j), x i, j+1 is the original image Pixel values at the image (i,j+1) coordinates:
对于对抗性损失使用以下交叉熵损失:For adversarial loss Use the following cross-entropy loss:
其中:pyadv()和py()分别是目标模型F(F指代通用机器模型的目标函数,例如vgg的目标函数F为fc8,由此可以得到对应1000类的概率输出)对标签yadv(对抗样本的类别)和y(原始图像的类别)的概率输出。Among them: p yadv () and p y () are the target model F respectively (F refers to the objective function of the general machine model, for example, the objective function F of vgg is fc8, from which the probability output corresponding to 1000 classes can be obtained) to the label yadv (the class of the adversarial example) and the probability output of y (the class of the original image).
将现实条件引入到增强示例的生成过程中,如下所示:Real-world conditions are introduced into the generation of augmented examples as follows:
其中:o是在物理世界中采样的随机背景图像,T是旋转、调整大小和颜色偏移的随机变换,是变换的集合;通过根据原始图像x和背景图像o,生成的增强样本对于人类观察者来说基本是合法的;where: o is a random background image sampled in the physical world, T is a random transformation for rotation, resizing and color shifting, is the set of transformations; by relying on the original image x and the background image o, the generated augmented samples are basically legal to a human observer;
目标背景重组增广使用分割算法Mask R-CNN将目标从背景中分割出来,使用插值算法将背景中空白部分补充像素,最后随机组合目标和背景,实现图片增广。The target background reorganization and augmentation uses the segmentation algorithm Mask R-CNN to segment the target from the background, uses the interpolation algorithm to supplement the blank part of the background with pixels, and finally randomly combines the target and the background to achieve image augmentation.
进一步的,多角度测试模块中的性能测试,包含不同角度:识别准确率Accuracy判别判断,识别损失值Loss判别判断以及蜕变关系判别;准确率Accuracy和损失值Loss的判断,都是用增广前后的模型输出的准确率Accuracy和识别损失值Loss相减得到的扩充前后识别准确率差值百分比Δacc和扩充前后识别损失差值百分比Δloss;Further, the performance test in the multi-angle test module includes different angles: the recognition accuracy rate Accuracy judgment, the recognition loss value Loss judgment judgment and the transformation relationship judgment; the accuracy rate and the loss value Loss judgment are both before and after augmentation. The difference between the accuracy rate of the model output Accuracy and the recognition loss value Loss is obtained by subtracting the recognition accuracy difference percentage Δacc before and after expansion, and the difference percentage of recognition loss before and after expansion Δloss;
蜕变测试定义为:Ci为原测试图像的被图像识别系统分类标签,Si为原测试图像的置信分数;Ci′为结合蜕变关系利用合成的新测试图像的分类标签,Si′为结合蜕变关系利用合成的新测试图像的置信分数,那么蜕变关系表述为:The transformation test is defined as: C i is the original test image is classified by the image recognition system, and S i is the original test image The confidence score of Synthesized new test image The classification label of , S i ′ is used in combination with the metamorphic relationship Synthesized new test image The confidence score of , then the transformation relationship is expressed as:
Ci=C′i andΔS=|Si-S′i|<c (7)C i =C' i andΔS=|S i -S' i |<c (7)
其中:c为超参数,0<c<100,c设置为50,ΔS为扩充前后置信分数的差。Where: c is a hyperparameter, 0<c<100, c is set to 50, and ΔS is the difference between the confidence scores before and after expansion.
进一步的,多角度测试模块中的可靠性测试为鲁棒性(certified robustness)测试,在原始图像x 满足置信度保证的条件下,在范数球半径R内能够免疫攻击:Further, the reliability test in the multi-angle test module is a robustness test. Under the condition that the original image x satisfies the confidence level guarantee, it can be immune to attacks within the radius R of the norm sphere:
其中:z()为损失函数,g()是待优化的目标函数,指任意,ε为引入的噪声,B(x;R)为噪声集合,R为范数球半径,R为一个无线接近0的值,x是原始图像;Among them: z() is the loss function, g() is the objective function to be optimized, Refers to any, ε is the noise introduced, B(x; R) is the noise set, R is the radius of the norm sphere, R is a wireless value close to 0, and x is the original image;
最后鲁棒性准确度(robacc)定义为:The final robustness accuracy (robacc) is defined as:
多角度测试模块中的安全性测试为模型不变性测试,选择一个随机图像,使用下面描述的四种方法之一选择一个像素的扰动,然后测量网络对该扰动的敏感度,第一种方法为“裁剪(Crop)”方法,在原始图像中随机选择一个正方形,并将该正方形的大小调整为224x224px,然后,我们将该正方形对角平移一个像素,以创建第二个图像,该图像通过平移单个像素来与第一个图像不同;第二种方法为“嵌入(Embedding)”方法,先缩小图像,使其最小尺寸为100px,同时保持纵横比,并将其嵌入到224x224px图像内的随机位置,同时用黑色(Black)像素填充图像的其余部分,然后将嵌入位置移位单个像素,再次创建两个相同的图像,直到移位单个像素;第三种方法中,先缩小图像,使其最小尺寸为100px,同时保持纵横比,并将其嵌入到224x224px图像内的随机位置,然后使用简单的修复算法(每个黑色像素被其邻域中的非黑色像素的加权平均所取代),第四个方法与第二个协议相同,先缩小图像,使其最小尺寸为100px,但我们不移动嵌入位置,而是保持嵌入位置不变,并将嵌入图像的大小更改单个像素(例如,从大小100x100px更改为大小101x101px像素)。The security test in the multi-angle test module is a model invariance test. A random image is selected, a perturbation of a pixel is selected using one of the four methods described below, and the sensitivity of the network to this perturbation is measured. The first method is The Crop method randomly selects a square in the original image and resize the square to 224x224px, then we translate the square by one pixel diagonally to create a second image, which is translated by A single pixel is different from the first image; the second method is the "Embedding" method, which first reduces the image to a minimum size of 100px, while maintaining the aspect ratio, and embeds it at a random position within the 224x224px image , while filling the rest of the image with black pixels, then shifting the embedding position by a single pixel, creating two identical images again, up to a single pixel shift; in the third method, the image is first scaled down to the smallest size is 100px while maintaining aspect ratio and embed it at random position inside 224x224px image, then use simple repair algorithm (each black pixel is replaced by weighted average of non-black pixels in its neighborhood), fourth The first method is the same as the second protocol, first shrink the image so that its minimum size is 100px, but instead of moving the embed position, we keep the embed position the same and change the size of the embed image by a single pixel (for example, from size 100x100px Change to size 101x101px).
进一步的,在安全性测试中,用两种方法测量灵敏度作为模型的不变性测试,第一个称之为 P(Top-1 Change),是网络的TOP-1预测在单像素扰动后发生变化的概率;第二称之为“平均绝对变化”(MAC),测量在顶层类(即在两个帧的第一帧中具有最高概率的类)的一个像素扰动之后,网络计算的概率的平均绝对值变化(即在两个帧的第一个帧中具有最高概率的类)。Further, in the security test, two methods are used to measure the sensitivity as the invariance test of the model. The first one is called P (Top-1 Change), which is the change of the TOP-1 prediction of the network after single-pixel perturbation. the probability of Absolute value change (ie the class with the highest probability in the first of the two frames).
进一步的,决策评估模块对所输入的测试结果进行分析,判断模型性能【Accuracy扩充后识别准确率,Loss扩充后识别损失,Δacc扩充前后识别准确率差,Δloss扩充前后识别损失差,CR 模型鲁棒性(用robacc来表征),ΔS扩充前后置信分数差,P(Top-1Change)网络的TOP-1预测在单像素扰动后发生变化的概率,MAC平均绝对变化)】,并给出详细的测试报告,当我们对多个识别模型进行性能对比的时候,大量单独的性能指标往往对用户来说过于繁杂,使用户难以作出合理的判断,因此,我们将不同指标对于识别系统的综合影响考虑到性能指标设计中,然后定义了一个综合性能指标CM(Composite Value)来反映不同识别系统的综合性能;公式如下:Further, the decision evaluation module analyzes the input test results and judges the performance of the model [accuracy after expansion, recognition loss after Loss expansion, poor recognition accuracy before and after Δacc expansion, poor recognition loss before and after Δloss expansion, and CR model is robust. Rodiness (characterized by robacc), difference in confidence scores before and after ΔS expansion, TOP-1 prediction probability of P(Top-1Change) network changes after single-pixel perturbation, MAC mean absolute change)], and gives detailed In the test report, when we compare the performance of multiple recognition models, a large number of individual performance indicators are often too complicated for users, making it difficult for users to make reasonable judgments. Therefore, we consider the comprehensive impact of different indicators on the recognition system. In the design of performance indicators, a comprehensive performance indicator CM (Composite Value) is defined to reflect the comprehensive performance of different identification systems; the formula is as follows:
其中:CMi代表第i个识别系统的综合性能值,ωj代表云服务第j个性能指标值的权重,max(Mj)代表多个识别系统中第j个性能指标的最大值,min(Mj)代表多个识别系统中第j个性能指标的最小值, Mij代表第i个识别系统的第j个性能指标值,N代表识别系统性能指标值的总数,通过利用公式 (2*max(Mj)-Mij)/(2*max(Mj)-min(Mj))对Mij的值进行标准化至[0,1]区间。可以看出,CM 的值越大,识别系统的综合性能越好。Among them: CM i represents the comprehensive performance value of the ith identification system, ω j represents the weight of the jth performance index value of the cloud service, max(M j ) represents the maximum value of the jth performance index in multiple identification systems, min (M j ) represents the minimum value of the j-th performance index in multiple recognition systems, M ij represents the j-th performance index value of the ith recognition system, and N represents the total number of performance index values of the recognition system. By using formula (2 *max(M j )-M ij )/(2*max(M j )-min(M j )) normalizes the value of M ij to the [0, 1] interval. It can be seen that the larger the value of CM, the better the comprehensive performance of the recognition system.
对于一些识别系统性能指标来说,例如Loss,P(Top-1 Change),MAC等,其值Mij越小, CM综合性能指标值越大,这些性能指标的值可以直接代入上述公式中,而对于识别准确率等指标,其值越大,说明识别性能更好,但是直接代入公式会造成CM值的降低,这并不符合预期,所以我们需要对这些性能指标的值进行处理,使用(1-Mij)来替代公式中的Mij值。For some identification system performance indicators, such as Loss, P(Top-1 Change), MAC, etc., the smaller the value M ij is, the larger the CM comprehensive performance index value is, and the values of these performance indicators can be directly substituted into the above formula, For indicators such as recognition accuracy, the larger the value, the better the recognition performance, but directly substituting the formula will cause the CM value to decrease, which is not in line with expectations, so we need to process the values of these performance indicators, using ( 1-M ij ) to replace the Mi ij value in the formula.
进一步的,决策评估模块最终会输出多角度测试模块测试的各项结果,并生成相应的测试报告表格,如表1-3所示。由于不同任务场景的要求不同,测试系统也会给出相应的建议。Further, the decision evaluation module will eventually output the results of the multi-angle test module test, and generate a corresponding test report form, as shown in Table 1-3. Due to the different requirements of different task scenarios, the test system will also give corresponding suggestions.
与现有技术相比,本发明具有以下有益效果:Compared with the prior art, the present invention has the following beneficial effects:
本发明实现了多类别影像测试数据生成和多角度完备的系统测试,最终完成医学影像识别系统的决策评估,未来应用前景广泛。The invention realizes multi-category image test data generation and multi-angle complete system test, finally completes the decision evaluation of the medical image recognition system, and has wide application prospects in the future.
附图说明Description of drawings
图1是本发明的框架流程图;Fig. 1 is the frame flow chart of the present invention;
图2是本发明中对抗增广流程图;Fig. 2 is a flow chart of confrontation augmentation in the present invention;
图3是本发明中目标背景重组增广流程图;Fig. 3 is a flow chart of target background reorganization and augmentation in the present invention;
图4是本发明中决策评估模块。Fig. 4 is a decision evaluation module in the present invention.
具体实施方式Detailed ways
下面结合实施例对本发明作更进一步的说明。The present invention will be further described below in conjunction with the examples.
一种针对医学影像识别系统的性能测试方法,如图1所示,性能测试方法中包括了:多类别影像测试数据生成模块、多角度测试模块和决策评估模块,所述多类别影像测试数据生成模块包括对抗样本生成网络和实体、背景重组方法;所述多角度测试模块包括性能测试、可靠性测试和安全性测试;决策评估模块,所述决策评估模块对所输入的测试结果进行分析,判断模型性能,并给出详细的测试报告;A performance test method for a medical image recognition system, as shown in Figure 1, the performance test method includes: a multi-category image test data generation module, a multi-angle test module and a decision evaluation module, the multi-category image test data generation module The module includes an adversarial sample generation network and entity, and a background reorganization method; the multi-angle testing module includes performance testing, reliability testing and security testing; a decision evaluation module, the decision evaluation module analyzes the input test results and judges Model performance, and give a detailed test report;
网络输入一组待分类识别的图片,首次将图片输入到所述多类别影像测试数据生成模块中,经过图像增广后输入模型进行分类,并将分类结果输入至多角度测试模块,多角度测试模块对模型的学习结果进行测试,并将结果传递到决策评估模块,决策评估模块对所输入的测试结果进行分析,判断模型性能,并给出详细的测试报告。The network inputs a set of pictures to be classified and recognized, and the pictures are input into the multi-category image test data generation module for the first time, and after image augmentation, the model is input for classification, and the classification results are input into the multi-angle test module, which is a multi-angle test module. The learning results of the model are tested, and the results are transmitted to the decision evaluation module. The decision evaluation module analyzes the input test results, judges the performance of the model, and gives a detailed test report.
对抗样本生成网络和实体、背景重组方法,考虑到医学影像的特性和生成的测试影像的视觉真实性,使用了对抗样本生成联合实体、背景重组方案。对抗增广使用多损失混合对抗伪装,该技术可以生成人类观察者看起来合法的新的增强图像,而不需要依赖大量的数据来训练生成网络。我们的目标是开发一种机制来生成具有自定义样式的扩展样本,利用样式变换技术实现图像增强,利用对抗攻击技术实现图像的隐蔽性。最终的多损失函数是对抗强度λ与对抗性损失的乘积、用于样式生成的样式损失用于保存源图像内容的内容损失和用于确保扩展样本的平滑度的平滑度损失的组合。The adversarial sample generation network and entity and background reorganization method, considering the characteristics of medical images and the visual authenticity of the generated test images, use the adversarial sample generation joint entity and background reorganization scheme. Adversarial augmentation uses multi-loss adversarial camouflage, a technique that can generate new augmented images that look legitimate to a human observer without relying on large amounts of data to train generative networks. Our goal is to develop a mechanism to generate extended samples with custom styles, image enhancement using style transformation techniques, and image concealment using adversarial attack techniques. The final multiple loss function is the adversarial strength λ and the adversarial loss The product of , the style loss for style generation Content loss for saving source image content and the smoothness loss used to ensure smoothness of the expanded samples The combination.
多损失函数表示为:Multiple loss functions Expressed as:
其中:λ表示对抗强度、表示对抗性损失、表示用于样式生成的样式损失、表示用于保存源图像内容的内容损失、表示用于确保扩展样本的平滑度的平滑度损失;Among them: λ represents the strength of confrontation, represents the adversarial loss, represents the style loss used for style generation, Indicates the content loss used to preserve the content of the source image, represents the smoothness loss used to ensure smoothness of the expanded samples;
如图2所示,显示了对抗增广方法的概述。用户定义现有图像、目标攻击区域和预期目标样式,在所需的区域中生成所需的样式,如图2右侧所示。为了使扩展样本对各种环境条件(包括照明、旋转等)具有鲁棒性,在每个步骤中向生成的扩展样本中添加附加的物理适应训练;As shown in Figure 2, an overview of adversarial augmentation methods is shown. The user defines the existing image, the target attack area, and the expected target style to generate the desired style in the desired area, as shown on the right side of Figure 2. To make the expanded samples robust to various environmental conditions (including lighting, rotation, etc.), additional physical adaptation training is added to the generated expanded samples at each step;
两个图像之间的样式距离通过两个图像在样式表示中的差异来定义,The style distance between two images is defined by the difference in the style representation of the two images,
其中:为特征距离,l是风格层特征,Sl是提取风格表示的风格层集,是风格样式的特征提取器,是从的一组样式层提取的深层特征的Gram矩阵,xs是风格参考图像,x′是生成的对抗样本;in: is the feature distance, l is the style layer feature, S l is the style layer set for extracting style representation, is a style-style feature extractor, From A set of Gram matrices of deep features extracted by the style layer, x s is the style reference image, and x′ is the generated adversarial example;
用于样式生成的样式损失所生成参考样式中的增强图像的内容与原始图像的内容非常不同;具体如下,style loss for style generation The content of the enhanced image in the generated reference style is very different from the content of the original image; as follows,
其中:是内容损失,t是内容层特征,ct是提取内容表示的内容层集,是内容层的特征提取器,x是原始图像,x′是生成的对抗样本;in: is the content loss, t is the content layer feature, c t is the content layer set from which the content representation is extracted, is the feature extractor of the content layer, x is the original image, and x′ is the generated adversarial example;
通过减少相邻像素之间的变化,提高增强图像的平滑度;对于增强图像,平滑度损失定义如下,The smoothness of the enhanced image is improved by reducing the variation between adjacent pixels; for the enhanced image, the smoothness loss is defined as follows,
其中:x′i,j为对抗样本(i,j)坐标处的像素值,xi+1,j为原始图像(i+1,j)坐标处的像素值,xi,j+1为原始图像(i,j+1)坐标处的像素值:Where: x′ i, j is the pixel value at the coordinate of the adversarial sample (i, j), x i+1, j is the pixel value at the coordinate of the original image (i+1, j), x i, j+1 is Pixel values at the (i,j+1) coordinates of the original image:
对于对抗性损失使用以下交叉熵损失:For adversarial loss Use the following cross-entropy loss:
其中:pyadv()和py()分别是目标模型F(F指代通用机器模型的目标函数,例如vgg的目标函数F为fc8,由此可以得到对应1000类的概率输出)对标签yadv(对抗样本的类别)和y(原始图像的类别)的概率输出。Among them: p yadv () and p y () are the target model F respectively (F refers to the objective function of the general machine model, for example, the objective function F of vgg is fc8, from which the probability output corresponding to 1000 classes can be obtained) to the label yadv (the class of the adversarial example) and the probability output of y (the class of the original image).
为了使对抗图像样本在现实世界中可实现,我们在生成增强样本的过程中对现实条件进行了建模。由于现实世界环境通常涉及条件波动,如视点移动、图像噪声和其他自然转换,因此我们使用一系列调整来适应这些不同的条件。特别地,我们使用了一种类似于期望过转换(EOT)的技术。我们的目标是提高扩充样本对不同物理条件的适应性。因此,我们考虑了用于模拟物理世界条件波动的变换,包括旋转、缩放、颜色偏移(以模拟光照变化)和随机背景。在这里,将现实条件引入到增强示例的生成过程中,如下所示:To make adversarial image samples achievable in the real world, we model real-world conditions during the generation of augmented samples. Since real-world environments often involve fluctuations in conditions such as viewpoint movement, image noise, and other natural transitions, we use a series of adjustments to accommodate these different conditions. In particular, we use a technique similar to expected over-transformation (EOT). Our goal is to improve the adaptability of augmented samples to different physical conditions. Therefore, we consider transformations used to simulate fluctuations in physical world conditions, including rotation, scaling, color shifting (to simulate lighting changes), and random backgrounds. Here, real-world conditions are introduced into the generation of augmented examples as follows:
其中:o是在物理世界中采样的随机背景图像,T是旋转、调整大小和颜色偏移的随机变换,是变换的集合;通过根据原始图像x和背景图像o,生成的增强样本对于人类观察者来说基本是合法的;where: o is a random background image sampled in the physical world, T is a random transformation for rotation, resizing and color shifting, is the set of transformations; by relying on the original image x and the background image o, the generated augmented samples are basically legal to a human observer;
目标背景重组增广使用分割算法Mask R-CNN将目标从背景中分割出来,使用插值算法将背景中空白部分补充像素,最后随机组合目标和背景,实现图片增广,总体方法框架如图3所示。Target background reorganization and augmentation The segmentation algorithm Mask R-CNN is used to segment the target from the background, and the interpolation algorithm is used to supplement the blank part of the background with pixels. Finally, the target and background are randomly combined to achieve image augmentation. The overall method framework is shown in Figure 3. Show.
多角度测试模块中的性能测试,包含不同角度:识别准确率Accuracy判别判断,识别损失值 Loss判别判断以及蜕变关系判别;准确率Accuracy和损失值Loss的判断,都是用增广前后的模型输出的准确率Accuracy和识别损失值Loss相减得到的扩充前后识别准确率差值百分比Δacc和扩充前后识别损失差值百分比Δloss;The performance test in the multi-angle test module includes different angles: recognition accuracy rate Accuracy judgment judgment, recognition loss value Loss judgment judgment and transformation relationship judgment; Accuracy rate and loss value Loss judgment are all output from the model before and after augmentation The difference between the recognition accuracy before and after expansion, Δacc, and the difference in recognition loss before and after expansion, Δloss, are obtained by subtracting the accuracy rate Accuracy and the recognition loss value Loss;
蜕变测试定义为:Ci为原测试图像的被图像识别系统分类标签,Si为原测试图像的置The transformation test is defined as: C i is the original test image is classified by the image recognition system, and S i is the original test image set
信分数;Ci′为结合蜕变关系利用合成的新测试图像的分类标签,Si′为结合蜕变关系利用合成的新测试图像的置信分数,那么蜕变关系表述为:trust score; C i ′ is used for the combined metamorphosis relationship Synthesized new test image The classification label of , S i ′ is used in combination with the metamorphic relationship Synthesized new test image The confidence score of , then the transformation relationship is expressed as:
Ci=C′ i andΔS=|Si-S′ i|<c (7)C i =C ′ i andΔS=|S i -S ′ i |<c (7)
其中:c为超参数,0<c<100,c设置为50,ΔS为扩充前后置信分数的差。Where: c is a hyperparameter, 0<c<100, c is set to 50, and ΔS is the difference between the confidence scores before and after expansion.
多角度测试模块中的可靠性测试为鲁棒性(certified robustness)测试,在原始图像x满足置信度保证的条件下,在范数球半径R内能够免疫攻击:The reliability test in the multi-angle test module is a certified robustness test. Under the condition that the original image x satisfies the confidence guarantee, it is immune to attacks within the radius R of the norm sphere:
其中:z()为损失函数,g()是待优化的目标函数,指任意,ε为引入的噪声,B(x;R)为噪声集合,R为范数球半径,R为一个无线接近0的值,x是原始图像;Among them: z() is the loss function, g() is the objective function to be optimized, Refers to any, ε is the noise introduced, B(x; R) is the noise set, R is the radius of the norm sphere, R is a wireless value close to 0, and x is the original image;
最后鲁棒性准确度(robacc)定义为:The final robustness accuracy (robacc) is defined as:
多角度测试模块中的安全性测试为模型不变性测试,选择一个随机图像,使用下面描述的四种方法之一选择一个像素的扰动,然后测量网络对该扰动的敏感度,第一种方法为“裁剪(Crop)”方法,在原始图像中随机选择一个正方形,并将该正方形的大小调整为224x224px,然后,我们将该正方形对角平移一个像素,以创建第二个图像,该图像通过平移单个像素来与第一个图像不同;第二种方法为“嵌入(Embedding)”方法,先缩小图像,使其最小尺寸为100px,同时保持纵横比,并将其嵌入到224x224px图像内的随机位置,同时用黑色(Black)像素填充图像的其余部分,然后将嵌入位置移位单个像素,再次创建两个相同的图像,直到移位单个像素;第三种方法中,先缩小图像,使其最小尺寸为100px,同时保持纵横比,并将其嵌入到224x224px图像内的随机位置,然后使用简单的修复算法(每个黑色像素被其邻域中的非黑色像素的加权平均所取代),第四个方法与第二个协议相同,先缩小图像,使其最小尺寸为100px,但我们不移动嵌入位置,而是保持嵌入位置不变,并将嵌入图像的大小更改单个像素(例如,从大小100x100px更改为大小101x101px像素)。The security test in the multi-angle test module is a model invariance test. A random image is selected, a perturbation of a pixel is selected using one of the four methods described below, and the sensitivity of the network to this perturbation is measured. The first method is The Crop method randomly selects a square in the original image and resize the square to 224x224px, then we translate the square by one pixel diagonally to create a second image, which is translated by A single pixel is different from the first image; the second method is the "Embedding" method, which first reduces the image to a minimum size of 100px, while maintaining the aspect ratio, and embeds it at a random position within the 224x224px image , while filling the rest of the image with black pixels, then shifting the embedding position by a single pixel, creating two identical images again, up to a single pixel shift; in the third method, the image is first scaled down to the smallest size is 100px while maintaining aspect ratio and embed it at random position inside 224x224px image, then use simple repair algorithm (each black pixel is replaced by weighted average of non-black pixels in its neighborhood), fourth The first method is the same as the second protocol, first shrink the image so that its minimum size is 100px, but instead of moving the embed position, we keep the embed position the same and change the size of the embed image by a single pixel (for example, from size 100x100px Change to size 101x101px).
在安全性测试中,用两种方法测量灵敏度作为模型的不变性测试,第一个称之为P(Top-1 Change),是网络的TOP-1预测在单像素扰动后发生变化的概率;第二称之为“平均绝对变化” (MAC),测量在顶层类(即在两个帧的第一帧中具有最高概率的类)的一个像素扰动之后,网络计算的概率的平均绝对值变化(即在两个帧的第一个帧中具有最高概率的类)。In the security test, two methods are used to measure the sensitivity as the invariance test of the model. The first one is called P (Top-1 Change), which is the probability that the TOP-1 prediction of the network will change after a single pixel perturbation; The second, called "Mean Absolute Change" (MAC), measures the mean absolute change in the probability computed by the network after one pixel perturbation of the top class (i.e. the class with the highest probability in the first of two frames) (i.e. the class with the highest probability in the first of the two frames).
如图4所示,决策评估模块对所输入的测试结果进行分析,判断模型性能【Accuracy扩充后识别准确率,Loss扩充后识别损失,Δacc扩充前后识别准确率差,Δloss扩充前后识别损失差,CR 模型鲁棒性(用robacc来表征),ΔS扩充前后置信分数差,P(Top-1 Change)网络的TOP-1预测在单像素扰动后发生变化的概率,MAC平均绝对变化)】,并给出详细的测试报告,当对多个识别模型进行性能对比的时候,大量单独的性能指标往往对用户来说过于繁杂,使用户难以作出合理的判断,因此,将不同指标对于识别系统的综合影响考虑到性能指标设计中,然后定义了一个综合性能指标CM(Composite Value)来反映不同识别系统的综合性能;公式如下:As shown in Figure 4, the decision evaluation module analyzes the input test results to judge the performance of the model [accuracy after expansion, recognition loss after Loss expansion, poor recognition accuracy before and after Δacc expansion, and poor recognition loss before and after Δloss expansion, The robustness of the CR model (characterized by robacc), the difference in confidence scores before and after ΔS expansion, the probability that the TOP-1 prediction of the P(Top-1 Change) network will change after a single-pixel perturbation, the MAC mean absolute change)], and A detailed test report is given. When comparing the performance of multiple recognition models, a large number of individual performance indicators are often too complicated for users, making it difficult for users to make reasonable judgments. Therefore, different indicators are used for the synthesis of the recognition system. The impact is taken into account in the design of performance indicators, and then a comprehensive performance indicator CM (Composite Value) is defined to reflect the comprehensive performance of different identification systems; the formula is as follows:
其中:CMi代表第i个识别系统的综合性能值,ωj代表云服务第j个性能指标值的权重,max(Mj)代表多个识别系统中第j个性能指标的最大值,min(Mj)代表多个识别系统中第j个性能指标的最小值, Mij代表第i个识别系统的第j个性能指标值,N代表识别系统性能指标值的总数,通过利用公式 (2*max(Mj)-Mij)/(2*max(Mj)-min(Mj))对Mij的值进行标准化至[0,1]区间。可以看出,CM 的值越大,识别系统的综合性能越好。Among them: CM i represents the comprehensive performance value of the ith identification system, ω j represents the weight of the jth performance index value of the cloud service, max(M j ) represents the maximum value of the jth performance index in multiple identification systems, min (M j ) represents the minimum value of the j-th performance index in multiple recognition systems, M ij represents the j-th performance index value of the ith recognition system, and N represents the total number of performance index values of the recognition system. By using formula (2 *max(M j )-M ij )/(2*max(M j )-min(M j )) normalizes the value of M ij to the [0, 1] interval. It can be seen that the larger the value of CM, the better the comprehensive performance of the recognition system.
对于一些识别系统性能指标来说,例如Loss,P(Top-1 Change),MAC等,其值Mij越小, CM综合性能指标值越大,这些性能指标的值可以直接代入上述公式中,而对于识别准确率等指标,其值越大,说明识别性能更好,但是直接代入公式会造成CM值的降低,这并不符合预期,所以我们需要对这些性能指标的值进行处理,使用(1-Mij)来替代公式中的Mij值。For some identification system performance indicators, such as Loss, P(Top-1 Change), MAC, etc., the smaller the value M ij is, the larger the CM comprehensive performance index value is, and the values of these performance indicators can be directly substituted into the above formula, For indicators such as recognition accuracy, the larger the value, the better the recognition performance, but directly substituting the formula will cause the CM value to decrease, which is not in line with expectations, so we need to process the values of these performance indicators, using ( 1-M ij ) to replace the Mi ij value in the formula.
决策评估模块最终会输出多角度测试模块测试的各项结果,并生成相应的测试报告表格,如表 1-3所示。由于不同任务场景的要求不同,测试系统也会给出相应的建议。The decision evaluation module will eventually output the results of the multi-angle test module test, and generate the corresponding test report form, as shown in Table 1-3. Due to the different requirements of different task scenarios, the test system will also give corresponding suggestions.
表1.性能指标报告Table 1. Performance Metrics Report
表2.安全性指标报告Table 2. Security Metrics Report
表3.模型稳定性指标报告以及模型综合性能报告Table 3. Model stability index report and model comprehensive performance report
以上所述仅是本发明的优选实施方式,应当指出:对于本技术领域的普通技术人员来说,在不脱离本发明原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本发明的保护范围。The above is only the preferred embodiment of the present invention, it should be pointed out that: for those skilled in the art, without departing from the principle of the present invention, several improvements and modifications can also be made, and these improvements and modifications are also It should be regarded as the protection scope of the present invention.
Claims (7)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011525218.0A CN112506797B (en) | 2020-12-22 | 2020-12-22 | A performance testing method for medical image recognition system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011525218.0A CN112506797B (en) | 2020-12-22 | 2020-12-22 | A performance testing method for medical image recognition system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112506797A CN112506797A (en) | 2021-03-16 |
CN112506797B true CN112506797B (en) | 2022-05-24 |
Family
ID=74923062
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011525218.0A Active CN112506797B (en) | 2020-12-22 | 2020-12-22 | A performance testing method for medical image recognition system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112506797B (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113268870B (en) * | 2021-05-19 | 2022-06-28 | 北京航空航天大学 | Monte Carlo-based image recognition reliability evaluation method under outdoor environment condition |
CN113486899B (en) * | 2021-05-26 | 2023-01-24 | 南开大学 | A Salient Object Detection Method Based on Complementary Branch Networks |
CN113780557B (en) * | 2021-11-11 | 2022-02-15 | 中南大学 | Anti-image attack method, device, product and medium based on immune theory |
KR20230103379A (en) | 2021-12-31 | 2023-07-07 | 삼성전자주식회사 | Method and apparatus for processing augmented reality |
CN114896134B (en) * | 2022-03-29 | 2024-07-30 | 中国科学院软件研究所 | Metamorphic testing method, device and equipment for target detection model |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008060022A1 (en) * | 2006-11-13 | 2008-05-22 | Electronics And Telecommunications Research Institute | System and method for evaluating and certifying image identifier |
CN111782529A (en) * | 2020-06-30 | 2020-10-16 | 平安国际智慧城市科技股份有限公司 | Test method and device for auxiliary diagnosis system, computer equipment and storage medium |
CN112052186A (en) * | 2020-10-10 | 2020-12-08 | 腾讯科技(深圳)有限公司 | Target detection method, device, equipment and storage medium |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10127146B2 (en) * | 2015-04-21 | 2018-11-13 | Cloudy Days, Inc. | Systems and methods to identify and classify performance bottlenecks in cloud based applications |
CN110008842A (en) * | 2019-03-09 | 2019-07-12 | 同济大学 | A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth |
CN110516695A (en) * | 2019-07-11 | 2019-11-29 | 南京航空航天大学 | Adversarial example generation method and system for medical image classification |
CN110458213A (en) * | 2019-07-29 | 2019-11-15 | 四川大学 | A Method for Robust Performance Evaluation of Classification Models |
CN111191660B (en) * | 2019-12-30 | 2022-07-22 | 浙江工业大学 | A multi-channel collaborative capsule network-based method for classifying pathological images of colon cancer |
CN111681210A (en) * | 2020-05-16 | 2020-09-18 | 浙江德尚韵兴医疗科技有限公司 | Method for identifying benign and malignant breast nodules by shear wave elastogram based on deep learning |
CN111753985B (en) * | 2020-06-28 | 2024-02-23 | 浙江工业大学 | Image deep learning model testing method and device based on neuron coverage rate |
-
2020
- 2020-12-22 CN CN202011525218.0A patent/CN112506797B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008060022A1 (en) * | 2006-11-13 | 2008-05-22 | Electronics And Telecommunications Research Institute | System and method for evaluating and certifying image identifier |
CN111782529A (en) * | 2020-06-30 | 2020-10-16 | 平安国际智慧城市科技股份有限公司 | Test method and device for auxiliary diagnosis system, computer equipment and storage medium |
CN112052186A (en) * | 2020-10-10 | 2020-12-08 | 腾讯科技(深圳)有限公司 | Target detection method, device, equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN112506797A (en) | 2021-03-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112506797B (en) | A performance testing method for medical image recognition system | |
CN111754596B (en) | Editing model generation method, device, equipment and medium for editing face image | |
Guo et al. | Fake face detection via adaptive manipulation traces extraction network | |
CN111723654B (en) | High-altitude parabolic detection method and device based on background modeling, YOLOv3 and self-optimization | |
Zhang et al. | Wildland forest fire smoke detection based on faster R-CNN using synthetic smoke images | |
CN109493308B (en) | Medical image synthesis and classification method for generating confrontation network based on condition multi-discrimination | |
CN110543846B (en) | Multi-pose face image obverse method based on generation countermeasure network | |
CN109492416A (en) | A kind of guard method of big data image and system based on safety zone | |
Tan et al. | Legitimate adversarial patches: Evading human eyes and detection models in the physical world | |
CN110348319A (en) | A kind of face method for anti-counterfeit merged based on face depth information and edge image | |
CN108537743A (en) | A kind of face-image Enhancement Method based on generation confrontation network | |
Liu et al. | FedForgery: generalized face forgery detection with residual federated learning | |
CN110533570A (en) | A kind of general steganography method based on deep learning | |
Peng et al. | CGR-GAN: CG facial image regeneration for antiforensics based on generative adversarial network | |
CN109977887A (en) | A kind of face identification method of anti-age interference | |
CN112784783B (en) | Pedestrian re-identification method based on virtual sample | |
CN112257741B (en) | A Generative Counterfeit Image Detection Method Based on Complex Neural Networks | |
CN110675353A (en) | Selective segmentation image synthesis method based on conditional generation countermeasure network | |
CN116563957B (en) | Face fake video detection method based on Fourier domain adaptation | |
CN111242868B (en) | Image enhancement method based on convolutional neural network in scotopic vision environment | |
Li et al. | High-capacity coverless image steganographic scheme based on image synthesis | |
Suryanto et al. | Active: Towards highly transferable 3d physical camouflage for universal and robust vehicle evasion | |
CN105138975A (en) | Human body complexion area segmentation method based on deep belief network | |
Mishra et al. | Text to image synthesis using residual GAN | |
CN111598144A (en) | Training method and device of image recognition model |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |