[go: up one dir, main page]

CN110807762B - An intelligent segmentation method of retinal blood vessel images based on GAN - Google Patents

An intelligent segmentation method of retinal blood vessel images based on GAN Download PDF

Info

Publication number
CN110807762B
CN110807762B CN201910884346.5A CN201910884346A CN110807762B CN 110807762 B CN110807762 B CN 110807762B CN 201910884346 A CN201910884346 A CN 201910884346A CN 110807762 B CN110807762 B CN 110807762B
Authority
CN
China
Prior art keywords
network
image
retinal
training
generator
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910884346.5A
Other languages
Chinese (zh)
Other versions
CN110807762A (en
Inventor
赵汉理
卢望龙
邱夏青
黄辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wenzhou University
Original Assignee
Wenzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wenzhou University filed Critical Wenzhou University
Priority to CN201910884346.5A priority Critical patent/CN110807762B/en
Publication of CN110807762A publication Critical patent/CN110807762A/en
Application granted granted Critical
Publication of CN110807762B publication Critical patent/CN110807762B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30041Eye; Retina; Ophthalmic
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30101Blood vessel; Artery; Vein; Vascular

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了一种基于GAN的视网膜血管图像智能分割方法,包括以下步骤:1、给定视网膜图像集,划分训练集和测试集;2、设计生成器网络G和判别器网络D,构建Adam优化器;3、将训练集输入到G;4、G生成血管分割图像;5、D对G生成的分割图像判别计算;6、对G和D参数更新;7、对G评估并获得最优模型G’,重复第3‑7步直至迭代结束;8、将视网膜图像输入G’生成血管分割图像。本发明使用大感受野网络模型对视网膜图像进行智能分割,得到最终的视网膜血管分割图像。本发明的网络模型具有较好的鲁棒性,所得到的血管分割图像含有较少的噪声,总体优于现有的视网膜血管图像分割方法。

Figure 201910884346

The invention discloses an intelligent segmentation method of retinal blood vessel images based on GAN. Optimizer; 3. Input the training set to G; 4. G generate blood vessel segmentation images; 5. D discriminate and calculate the segmented images generated by G; 6. Update the parameters of G and D; 7. Evaluate G and obtain the optimal Model G', repeat steps 3-7 until the iteration ends; 8. Input the retinal image into G' to generate a blood vessel segmentation image. The invention uses the large receptive field network model to intelligently segment the retinal image to obtain the final retinal blood vessel segmented image. The network model of the invention has better robustness, and the obtained blood vessel segmentation image contains less noise, which is generally better than the existing retina blood vessel image segmentation method.

Figure 201910884346

Description

Intelligent retinal blood vessel image segmentation method based on GAN
Technical Field
The invention belongs to the field of intelligent segmentation of retinal vessel images, and particularly relates to an intelligent segmentation method of a retinal vessel image based on GAN, which better solves the problem that the accuracy rate of the conventional retinal vessel image segmentation method is low.
Background
In clinical medicine, doctors often analyze ophthalmology and some systemic diseases, such as diabetes, glaucoma, hypertension, cardiovascular and cerebrovascular diseases, and the like by observing morphological characteristics of retinal images. The occurrence of these diseases generally affects the human retinal morphology, such as information affecting the number, branching, width, and angle of retinal blood vessels. Therefore, the realization of the segmentation of the color retinal image becomes an important condition for disease judgment of ophthalmologists, but the segmentation of the color retinal image by using the manual method is time-consuming and labor-consuming, and the segmentation result is also influenced by the experience of operators and the segmentation technology. The method has strong subjectivity and low repeatability, so that the method has important medical research significance and application value for realizing the intelligent and accurate segmentation of the retinal vessel image. With the continuous development of computer-aided diagnosis systems in medicine, intelligent segmentation of retinas also becomes a research hotspot at present.
Disclosure of Invention
The invention provides an intelligent retinal vessel image segmentation method based on GAN (generic object network), aiming at the problems of strong subjectivity and low efficiency in manual retinal vessel image segmentation and the problem of low segmentation accuracy in the existing method for performing retinal vessel segmentation by using a supervision method.
In order to solve the technical problems in the prior art, the technical scheme of the invention is as follows: a retinal vessel image intelligent segmentation method based on GAN comprises the following steps:
in step S1, a retinal image sample set is given, containing a sample pair of a retinal image and a reference blood vessel segmentation image, defined herein as (a, b); defining a retina image corpus C { (a)i,bi)|i∈[1,R]R denotes the total number of samples, i denotes the number of samples, a denotes the retinal image, and b denotes the reference blood vessel segmentation image. Copying and dividing a retina image sample set into a retina image training set E { (a)i,bi)|i∈[1,M]And a retinal image test set F { (a) }i,bi)|i∈[1,N]Where N + M ═ R, and M and N respectively denote the corresponding number of samples.
Step S2, constructing a retina intelligent segmentation network model based on GAN, wherein the network model comprises a generator network G and a discriminator network D, and constructing an Adam optimizer to assist the network training to quickly converge:
the overall architecture of the generator network G includes two parts, a contracting path (contracting path) and an expanding path (expanding path). In order to utilize the characteristic diagram information in the network training process to a greater extent, the characteristic diagram extracted and processed on the network contraction path is spliced with the characteristic diagram in the expansion path with the same size in the process of upsampling. In addition, a cavity convolution structure is introduced into the bottom layer of network downsampling, and the structure can increase the receptive field of the generated network, so that the network can better grasp the global characteristics of the retinal vessel image, and accurate segmentation of the retinal vessel image is realized. The generator network G performs 4 downsampling, 4 upsampling, and 3 feature concatenating operations in total, and the selected feature map is the feature map after downsampling, so that the generator network performs only 3 feature map concatenating operations although downsampling is performed 4 times. The down-sampling operation used in the generator network G is performed using a convolution operation with a convolution kernel size of 3x3 steps of 2.
The discriminator network D is a deep convolutional neural network, and its main role is to judge whether the input blood vessel segmented image is a reference blood vessel segmented image or a blood vessel segmented image generated by the generator network G. A residual block ResBlock structure is also used in the discrimination network, the structure can prevent the over-fitting of the network while increasing the number of network layers, and the problem of difficult training is solved, so that the network can better capture image characteristics, and the network can be converged more quickly. In the discriminator network D, the convolution kernel size used is 3 × 3, and then the downsampling operation is performed using the maximum pooling layer maxPooling operation with step size of 2, highlighting the main features in the feature map.
It is noted that the generator countermeasure network is composed of a generator network G and a discriminator network D. The main process is that the generator network G continuously fits the distribution of the retina training set E, and inputs the retina training sample pair (a)i,bi) The generator network G generates a vessel segmentation image ziAnd obtaining a retina sample pair (a) generated by the generatori,zi). The discriminator network D simultaneously and respectively aligns the E sample pairs (a) of the retina training seti,bi) And the generated retina sample pair (a)i,zi) Discrimination is made wherein i is 1,2,3, …, M, each giving [0,1]The discrimination confidence q between the samples represents the summary of the sample pair as the retina training set E sample pairAnd (4) rate. Vessel segmentation sample z generated by loss functioniAnd a reference blood vessel segmentation sample biIn between, in preparation for further back propagation.
Constructing an Adam optimizer to assist network training, and setting an initial learning rate of 0.0002, beta1=0.5,β2The learning rate can be intelligently adjusted during the training process, so that the network can be converged quickly.
Step S3, loading the retina training set E into the computer memory as input, and setting the retina training set E { (a)i,bi)|i∈[1,M]And (4) randomly scrambling to prepare for the next training stage.
Steps S4, S5, and S6 are the main training phase for generating the countermeasure network, and the game problem of the discriminator network D and the generator network G in the generation countermeasure network can be considered as a maximum minimization problem, and the two networks understand the relationship of image mapping therein by learning the features of the retinal image to the blood vessel segmentation image. The objective function is shown in equation 1:
Figure BDA0002206830800000031
at each iteration, a pair of image pairs (a) in the retina training set E is extractedi,bi),aiRepresenting a retinal image, biA reference blood vessel segmentation image is shown, where i is 1,2,3, …, M.
In step S4, a is input from the generatoriGenerating a corresponding blood vessel segmentation image G (a)i) I.e. ziThe generator network G tries to minimize the objective function LcGAN(G, D), in order to make the output of the final objective function as small as possible, a vessel segmentation image z is generatediThe image b is segmented as much as possible in the image style, the vascular structure and the reference retinaiAs similar as possible.
In step S5, the discriminator network D attempts to discriminate the distribution of the retina training set E and the distribution of the retina training composite set E' so as to maximizeThe discriminant network D simultaneously and respectively processes the sample pairs (a) of the retina training set Ei,bi) And the sample pairs of the retinal training composition set E' (a)i,zi) Discrimination is made wherein i is 1,2,3, …, M, each giving [0,1]The discrimination confidence q between the samples q represents that the sample pair is a retina training set E sample pair (a)i,bi) The probability of (c).
Finally, the game equilibrium points of the discriminator network D and the generator network G are 'Nash equilibrium points', that is, the discriminator network D cannot judge that the input image sample pair is the sample pair (a) of the retina training set Ei,bi) Or the sample pair of the retina training synthetic set E' (a)i,zi) For a given retina, sample pairs (a) of the synthetic set E' are trainedi,zi) D confidence q for each output is 0.5. At this time, the distribution of the segmented image generated by the generator network G is fitted with the distribution of the reference blood vessel segmented image, so that the accurate mapping relation from the retina image to the blood vessel segmented image is learned, and the generated segmented image is the target image required by people. The process of the game can be considered as a maximum minimization process, which can be expressed as:
Figure BDA0002206830800000032
since the segmentation from the retina image into the blood vessel segmentation image is essentially a classification prediction of "black or white" for each pixel, which is actually a pixel-to-pixel classification task, the present invention additionally uses a class-two classification cross entropy loss in the generator network G to penalize the distance between the generated blood vessel segmentation image and the reference blood vessel segmentation image, so that the generated blood vessel segmentation image is more approximate to the reference blood vessel segmentation image. The class-two class cross entropy loss function is defined as follows:
Figure BDA0002206830800000041
where a is the retinal image, b is the reference vessel segmentation map, and G is the corresponding generator network G. In step S6, based on the loss functions given by equation (1) and equation (3), a total loss value at the current iteration number can be calculated. In order to minimize the loss value, a gradient value of each parameter in each step can be obtained by using a computation graph, and the whole function is close to a minimum value point by using a gradient updating method, so that the aim of fitting is fulfilled. The corresponding parameter updating formula is as follows:
Figure BDA0002206830800000042
wherein theta istParameters representing the t-th component in the generator network G and the discriminator network D, η represents the learning rate in the hyper-parameters,
Figure BDA0002206830800000043
representing the gradient of the corresponding parameter.
In step S7, the generator is evaluated using the retinal image test sample, and the optimal model parameters are retained as follows: at the end of the training phase, inputting the retina test set F into a generator network G, and generating a retina test synthesis set F' by using the retina test set F sample pairs (a)i,bi) Reference vessel segmentation map b in (1)iAnd the generated retina sample pair (a)i,zi) Generating a segmentation map z of (1)iA pixel-by-pixel alignment is performed, where i is 1,2,3, …, N, and each pixel is classified as a vascular point and a non-vascular point. In order to perform a performance test on the current generator network G, quantitative analysis needs to be objectively performed by a performance index. Indexes such as Accuracy (Accracy, Acc), Specificity (Sp), Sensitivity (Se), Dice coefficient, F-measure, Area (AUC) formed by a receiver operating characteristic curve (ROC) and a coordinate axis, and area (mAP) formed by an Accuracy-recall rate curve (PR curve) and the coordinate axis are adopted to measure the effectiveness of the model.
The AUC is more used for the performance measurement of medical image processing, and the closer the value of the AUC is to 1, the better the segmentation effect is.
Figure BDA0002206830800000044
Figure BDA0002206830800000045
Figure BDA0002206830800000046
Figure BDA0002206830800000051
Figure BDA0002206830800000052
Figure BDA0002206830800000053
Wherein, TP (true positive) is true positive and represents the number of correctly segmented blood vessels; TN (true negative) represents the number of correctly segmented non-blood vessels, namely background pixels; FP (false positive), that is, the number of pixel points of the blood vessel which is wrongly divided into non-blood vessels; FN (false negative), i.e., the number of pixels in the non-blood vessel is erroneously classified as blood vessel. TP + FN + FP + TN is the total number of pixel points in the region of interest in the image.
Since the above evaluation index depends on the threshold value of the output result, ROC data can be plotted by changing the true positive score (Sensitivity) and the false positive score (Specificity), and AUC is the area under the ROC curve. All evaluation indices are tests that are considered on all pixels within a mask, which represents the retinal optic disc region. And screening the indexes, and selecting the model with the largest indexes of Acc, Se, Sp, Precision, Recall and F-measure as the optimal model. And finally, judging at a parameter updating end stage, judging whether the training iteration number reaches the maximum iteration number, if so, ending the training stage to obtain an optimal generator network G 'and an optimal discriminator network D', and entering the next step. Otherwise, step S3 is entered for continuous loop iteration training.
The training image and the test image input in the training and testing process are the whole image. According to the invention, an Adam optimization method is adopted to carry out optimization training on the loss function, so that the final parameters for generating the confrontation network model are obtained, and the parameters can be continuously used in the subsequent retina segmentation task after being stored.
In step S8: a retina image sample set F1={ai|i∈[1,RF]A, taking each retinal image aiAs input in the optimal generator network G, the corresponding vessel segmentation image z is outputi,i=1,2,3,…,RFIn the formula, RFThe sample number of the retina image sample set is represented, and the final segmentation image has better accuracy.
The invention provides an intelligent retinal vessel image segmentation method based on a generation countermeasure network, which has the beneficial effects that:
the network is mainly characterized by being based on a countermeasure training mechanism, and has a larger receptive field, so that the global information of the image can be captured well. Compared with other segmentation networks, the network has deeper network layers and can better capture and utilize abstract features of images. The method achieves the advanced effects in the aspects of accuracy, sensitivity and specificity. In addition, good segmentation effect can be achieved in a blood vessel region and a lesion region with low contrast, the method achieves high precision and good robustness of retinal blood vessel segmentation, and has good value and prospect in practical application.
Drawings
FIG. 1 is a general flow diagram of the present invention;
FIG. 2 is an overall framework of the invention;
FIG. 3 is a generator architecture diagram of the present invention;
FIG. 4 is a block diagram of a Residual block used in the present invention;
FIG. 5 is a block diagram of a scaled residual block of the present invention;
FIG. 6 is a discriminator architecture diagram of the present invention;
fig. 7 is a diagram of the final segmentation effect of the present invention.
Detailed Description
For completeness and clarity of description of technical solutions in the embodiments of the present invention, the following detailed description will be further developed with reference to the accompanying drawings in the embodiments of the present invention. It should be understood that the specific embodiments described herein are merely illustrative of the invention and do not limit the invention.
As shown in fig. 1, the present invention provides a technical solution: a retinal vessel image intelligent segmentation method based on GAN comprises the following steps:
step S1: given a sample set of retinal images, a sample pair comprising a retinal image and a reference vessel segmentation image, defined herein as (a, b); defining a retina image corpus C { (a)i,bi)|i∈[1,R]R denotes the total number of samples, i denotes a sample index, a denotes a retina image, and b denotes a reference blood vessel segmentation image. Copying and dividing a retina image sample set into a retina image training set E { (a)i,bi)|i∈[1,M]And a retinal image test set F { (a) }i,bi)|i∈[1,N]Where N + M ═ R, and M and N respectively denote the corresponding number of samples.
In the invention, the main steps comprise data division, and generation of the antagonistic loss is carried out by combining the constructed generator network G and the discriminator network D, finally the generator network G and the discriminator network D reach the optimal balance point, and the accurate segmentation of the input retina image by using the generator network G is realized, and the relationship between the overall network structure and the loss function is shown in FIG. 2.
Step S2: combining characteristic up-down sampling, a residual block structure and cavity convolution operation to obtain a generator network G; designing a discriminator network D, and adding a discriminator network D with a large receptive field and an improved residual structure into the discriminator network D; the combination of the generator network G and the arbiter network D is finally referred to as the generation countermeasure network. And initializing network parameters of the generator network G and the discriminator network D by using an Adam optimization method in a Pythrch frame to obtain initial parameters of the generator network and the discriminator network, and setting related training hyper-parameters for training and optimizing the network model.
The generator network G is specifically constructed by combining the advantages of feature up-down sampling and a ResBlock structure, so that the network combines the shallow feature of the image while up-sampling, thereby more comprehensively utilizing the feature map information and preventing the network degradation problem in a deep network. In addition, the generation network uses the hole convolution operation, and the receptive field of the generation network is increased while the number of network parameters is not increased. The sizes of convolution kernels in the generated network are all 3x3, after each next sample operation, the output feature map is input into 2 residual blocks ResBlock to carry out jump connection operation, and the feature map obtained after 4 times of downsampling is subjected to hole convolution operation with the hole rates of 5 and 3 respectively to increase the receptive field of the network. In the process of up-sampling, the features extracted by the shallow layer network are spliced for operation, the network finally uses convolution operation of 1x1 to ensure that the number of color channels output by the network is the same as the number of input color channels, the network structure is shown in fig. 3, and an upward arrow in the legend represents deconvolution operation, BatchNormal operation and Relu activation operation, wherein the convolution kernel size is 3x3 and the step size is 2; "downward arrow" indicates the convolution operation, BatchNormalization operation, Relu activation operation with a convolution kernel size of 3 × 3 step size of 2; "thin arrow to the right" indicates the convolution operation, BatchNormalization operation, Relu activation operation with a convolution kernel size of 3 × 3 step size of 1; "Thick arrows to the right" indicate convolution operation, BatchNormalization operation, Relu activation operation with convolution kernel size 1 × 1 step size 1; the "feature map" represents an output image after a corresponding convolution operation, i.e., is a feature map; "dotted arrows" indicate that the corresponding output feature maps are subjected to a splicing operation, that is, based on the dimension of the color channel, the number of the color channels is spliced, for example, the feature map with the size of (m, h, w) and the feature map with the size of (n, h, w) are subjected to a splicing operation, and the obtained feature map has the size of (m + n, h, w); the "residual block" refers to an operation based on a residual block network, as shown in fig. 4, the arrows in the diagram do not contain any operation meaning, and only represent the operation flow, and the residual block respectively contains two convolution operations with convolution kernel size of 3 × 3 and step size of 1, a batch normalization operation, and a Relu activation operation, and when the final output is performed, the input original feature map and the output feature map of the last convolution are added and summed to obtain and output the feature map with the same number, height, and width of color channels as the input feature map; "hole residual block" is based on the residual block, except that two convolution kernels are respectively set to different hole rates, other processes are all the same as the "residual block", and a schematic diagram is shown in fig. 5.
The overall architecture of the generator network G includes two parts, a contracting path (contracting path) and an expanding path (expanding path). The contraction path represents an operation path based on the U shape, namely an operation path which is input from an image, subjected to four times of down-sampling, further subjected to a hole residual block, and then subjected to 4 times of up-sampling output; the extended path indicates a path of the splicing operation other than the path of the U-shape. In order to utilize the feature information in the network to a greater extent, the feature map of the features extracted on the network contraction path after up-sampling is spliced with the feature map in the expansion path with the same size.
In the generator network G, 4 downsampling operations, 4 upsampling operations and 3 feature splicing operations are performed in total, and the selected and spliced feature map is the feature map after downsampling. And the downsampling operations used in the generation network are all performed using convolution operations with a convolution kernel size of 3x3 steps of 2.
The construction of the discriminator network D is specifically that a ResBlock structure is introduced into the discriminator network D to prevent the degradation problem of the deep network, and the used hopping connection structure still enables the network to optimize learning under the condition of increasing the network depth, and simultaneously prevents the network degradation problem. The size of each layer of convolution kernel in the network is judged to be 3x3, the network is downsampled by carrying out Max boosting operation for 4 times in total, the network finally uses a full connection layer to classify vectors, and whether the image input into the judgment network is from a real image or an image generated by a generator is judged. The structure of the discrimination network is shown in fig. 6, wherein in the legend, "Scalar" represents a Scalar, that is, the output value of the final discriminator network D, and is between [0,1], which represents the confidence of the final discriminator network in the authenticity judgment of the input image; "feature map" means the output image after the convolution operation; "thin arrow to the right" indicates the convolution operation with a convolution kernel size of 3 × 3 steps of size 1, the batch normalization operation, the Relu activation operation; "Thick arrows to the right" indicate convolution operations, BatchNormalization operations, Relu activation operations with convolution kernel sizes of 3 × 3 steps of size 2; "thin dashed arrow to the right" represents a flattening operation that stretches the feature map of the multi-color channel into a one-dimensional vector; the "residual block" here is the same as the "residual block" in the generator network G; "maximum pooling layer" means the maximum downsampling operation with an operation kernel size of 2 x 2 steps of size 2, with the output image size being 1/2 of the original image in both height and width; "global average pooling" means a downsampling operation that operates on the average of kernel size to image size, with the output result such that the image sizes are averaged over all, (c, h, w) size images are "global average pooled" to (c, 1, 1), where c is the number of color channels, h is the height of the feature map, and w is the width of the feature map; "thick dotted arrow to the right" indicates a full join operation, i.e., the target vector is transformed into a matrix to obtain a result vector.
It is noted that the numbers in the upper left corner of all feature maps in the generator network (as shown in fig. 3) and the discriminator network (as shown in fig. 6) represent the number of color channels, corresponding to the number of color channels output after each pass of the "arrow operation". And the 'K' at the lower left corner of the feature map represents the size of the original image, and the size of the feature map is changed into 'K/p' after multiple times of down-sampling, which represents that the adopted size is K divided by the corresponding multiple p.
Meanwhile, an Adam optimization method is adopted in the Pythrch framework, and the hyper-parameters in the Pythrch framework are optimized in the training process, so that the initial parameters of the generator network G and the discriminator network D are assisted.
Step S3: loading the retina training set E into a memory, randomly disorganizing, and extracting a pair of retina training samples (a) for each trainingi,bi) Including the retinal image and the reference blood vessel segmentation image, to prepare for the next training.
Step S4: extracting a pair of samples (a) from a retina training set Ei,bi) I is 1,2,3, …, M. The retina image aiAnd loading and inputting the image into a generator network G, wherein the corresponding image size is (3 multiplied by h multiplied by w), 3 represents the number of color channels, the corresponding color channel d belongs to { red, green and blue, h represents the height of a single picture, and w represents the width of the single picture. Through the layer-by-layer calculation of the network, the generated retina blood vessel segmentation graph z is finally obtainediThe image size is (1 × h × w), and the corresponding gray-scale map is a single color channel, i.e., the image is represented by the gray-scale value size of the image according to the degree of the blood vessel in the image.
Step S5: the discriminant network D attempts to maximize the loss function by distinguishing the distribution of the retina training set E from the distribution of the retina training composite set E', and the discriminant network D simultaneously and separately processes the pairs of retina training set E samples (a)i,bi) And retina training synthetic set E' sample pairs (a)i,zi) Discrimination is made wherein i is 1,2,3, …, M, each giving [0,1]The discrimination confidence q between the samples q represents the probability that the sample pair is a retina training set E sample pair.
Step S6: and calculating the error between the generated retina training synthetic set E' and the retina training set E through loss function calculation to obtain a loss value. And performing backward propagation by using the obtained loss value, and respectively performing network parameter adjustment on the discriminator network D and the generator network G. And according to the given loss function, calculating the gradient of the parameters in the generator network G and the discriminator network D by using a chain type derivative method, and updating the corresponding parameters by using a random gradient descent method. The corresponding parameter updating formula is as follows:
Figure BDA0002206830800000111
wherein theta istParameters representing the t-th component in the generator network G and the discriminator network D, η represents the learning rate in the hyper-parameters,
Figure BDA0002206830800000112
representing the gradient of the corresponding parameter.
Step S7: the generators were evaluated using the retina test set F, with the optimal model parameters retained. And meanwhile, judging at the parameter updating end stage, judging whether the training iteration number reaches the maximum iteration number, if so, ending the training stage, and entering the next step. Otherwise, the training is continued, and the step S3 is continued to continue the loop iteration training.
At the end of the model training phase, inputting the retina test set F into a generator network G, and generating a retina test synthesis set F' by using the retina test set F sample pairs (a)i,bi) Reference vessel segmentation map b in (1)iAnd the generated retina sample pair (a)i,zi) Generating a segmentation map z of (1)iA pixel-by-pixel alignment is performed, where i is 1,2,3, …, N, and each pixel is classified as a vascular point and a non-vascular point. In order to perform a performance test on the current generator network G, quantitative analysis needs to be objectively performed by a performance index. Indexes such as Accuracy (Accracy, Acc), Specificity (Sp), Sensitivity (Se), Dice coefficient, F-measure, Area (AUC) formed by a receiver operating characteristic curve (ROC) and a coordinate axis, and area (mAP) formed by an Accuracy-recall rate curve (PR curve) and a coordinate axis are adopted to measure the text modelThe effectiveness of the model.
The AUC is more used for the performance measurement of medical image processing, and the closer the value of the AUC is to 1, the better the segmentation effect is.
Figure BDA0002206830800000121
Figure BDA0002206830800000122
Figure BDA0002206830800000123
Figure BDA0002206830800000124
Figure BDA0002206830800000125
Figure BDA0002206830800000126
Wherein, TP (true positive) is true positive and represents the number of correctly segmented blood vessels; TN (true negative) represents the number of correctly segmented non-blood vessels, namely background pixels; FP (false positive), that is, the number of pixel points of the blood vessel which is wrongly divided into non-blood vessels; FN (false negative), i.e., the number of pixels in the non-blood vessel is erroneously classified as blood vessel. TP + FN + FP + TN is the total number of pixel points in the region of interest in the image.
Since the above evaluation index depends on the threshold value of the output result, ROC data can be plotted by changing the true positive score (Sensitivity) and the false positive score (Specificity), and AUC is the area under the ROC curve. All evaluation indices are tests that are considered on all pixels within a mask, which represents the retinal optic disc region. And screening the indexes, and selecting the model with the largest indexes of Acc, Se, Sp, Precision, Recall and F-measure as the optimal model. And finally, judging at a parameter updating end stage, judging whether the training iteration number reaches the maximum iteration number, if so, ending the training stage to obtain an optimal generator network G 'and an optimal discriminator network D', and entering the next step. Otherwise, step S3 is entered for continuous loop iteration training.
Step S8: and testing the test sample image by using the trained generator network G, inputting an original retinal image, and correspondingly outputting a retinal blood vessel segmentation image.
A retina image sample set F1={ai|i∈[1,RF]A, taking each retinal image aiAs input in the optimal generator network G, the corresponding vessel segmentation image z is outputi,i=1,2,3,…,RFWhere RF represents the number of samples of the retinal image sample set. The segmentation effect of the present invention on a sample set of retinal images is shown in fig. 7.
In summary, the invention adopts a GAN-based retinal vessel image intelligent segmentation method, and the generated network uses a residual error structure and a hole convolution operation, so that the network increases the receptive field of the network without introducing additional parameters, the network captures the characteristics of the image more comprehensively, and the retinal segmentation task can be completed better. In addition, a feature stitching operation in the expanded path of the network is generated, which enables the network to better exploit the shallow and deep features of the image for segmentation tasks. In the discrimination network, the structure of a residual error module is added, so that the problem of network degradation caused by the deepening of the network layer number is avoided, the discrimination capability of a deep network is better utilized, and the supervision capability of the network is enhanced.
It will be appreciated by persons skilled in the art that the invention is not limited to details of the foregoing embodiments, and that the invention can be embodied in other specific forms without departing from the spirit or scope of the invention. In addition, various modifications and alterations of this invention may be made by those skilled in the art without departing from the spirit and scope of this invention, and such modifications and alterations should also be viewed as being within the scope of this invention.

Claims (1)

1.一种基于GAN的视网膜血管图像智能分割方法,其特征在于,所述方法包括以下步骤:1. a kind of retinal blood vessel image intelligent segmentation method based on GAN, is characterized in that, described method comprises the following steps: 步骤S1:给定视网膜图像样本集,样本集包含视网膜图像和基准血管分割图像的样本对,视网膜图像样本集记为C={(ai,bi)|i∈[1,R]},式中,a表示视网膜图像,b表示基准血管分割图像,R表示样本数量,i表示样本下标;将视网膜图像全集C拷贝并划分成视网膜图像训练集E={(ai,bi)|i∈[1,M]}和视网膜图像测试集F={(ai,bi)|i∈[1,N]},其中N+M=R,且M和N分别表示划分后的样本数量;Step S1: Given a retinal image sample set, the sample set includes a sample pair of retinal image and reference blood vessel segmentation image, and the retinal image sample set is denoted as C={(a i ,b i )|i∈[1,R]}, In the formula, a represents the retinal image, b represents the reference blood vessel segmentation image, R represents the number of samples, and i represents the sample subscript; the retinal image complete set C is copied and divided into retinal image training set E={(a i ,b i )| i∈[1,M]} and retinal image test set F={(a i ,b i )|i∈[1,N]}, where N+M=R, and M and N represent the divided samples, respectively quantity; 步骤S2:构建生成器网络G,其中包含残差块结构和空洞卷积操作;设计判别器网络D,其中包含残差结构,得到大感受野的判别器网络D;最终将生成器网络G和判别器网络D的组合称为生成对抗网络;在Pytorch框架中使用Adam优化方法,对生成器网络G和判别器网络D的网络参数进行初始化,得到生成器网络G和判别器网络D的初始参数,并设定相关训练超参数,用于网络模型的训练优化;Step S2: Build a generator network G, which includes a residual block structure and a hole convolution operation; design a discriminator network D, which includes a residual structure, and obtain a discriminator network D with a large receptive field; Finally, the generator network G and The combination of the discriminator network D is called a generative adversarial network; the Adam optimization method is used in the Pytorch framework to initialize the network parameters of the generator network G and the discriminator network D, and the initial parameters of the generator network G and the discriminator network D are obtained. , and set the relevant training hyperparameters for training optimization of the network model; 步骤S3:将视网膜训练集E作为输入,载入到计算机内存中,并将视网膜训练集E={(ai,bi)|i∈[1,M]}随机打乱,为接下来的训练阶段做准备;Step S3: take the retina training set E as input, load it into the computer memory, and randomly scramble the retina training set E={(a i ,b i )|i∈[1,M]} for the next step. Prepare for the training phase; 步骤S4:生成器网络G将视网膜训练集E作为输入,并通过网络逐层计算输出,生成对应的视网膜训练合成集E'={(ai,zi)|i∈[1,M]},其中E'中的ai和E中的ai相同,皆为视网膜图像;而E'中的zi表示生成器网络G生成的血管分割图,E中的bi表示基准的血管分割图;Step S4: The generator network G takes the retina training set E as input, and calculates the output layer by layer through the network to generate the corresponding retina training synthesis set E'={(a i ,z i )|i∈[1,M]} , where a i in E' is the same as a i in E, both are retinal images; and zi in E' represents the blood vessel segmentation map generated by the generator network G, and b i in E represents the benchmark blood vessel segmentation map ; 步骤S5:判别器网络D对生成器网络G生成的视网膜训练合成集E'和视网膜训练集E中的图像样本分别逐一进行判定,对每对生成的视网膜样本对(ai,zi)和视网膜训练样本对(ai,bi)给出判定的置信度q,i=1,2,3,...,M,q范围在[0,1]之间;Step S5: The discriminator network D judges the image samples in the retinal training synthetic set E' and the retinal training set E generated by the generator network G one by one, respectively. The retina training sample pair (a i , b i ) gives the confidence q of the decision, i=1, 2, 3,...,M, and the range of q is between [0, 1]; 步骤S6:通过损失函数计算,计算生成的视网膜训练合成集E'和视网膜训练集E之间的误差,得出损失值;利用得出的损失值进行反向传播,对判别器网络D和生成器网络G分别进行网络参数调整;Step S6: Calculate the error between the generated retinal training set E' and the retinal training set E by calculating the loss function, and obtain a loss value; use the obtained loss value to perform backpropagation, and compare the discriminator network D and generation. The network parameters of the device network G are adjusted respectively; 步骤S7:利用视网膜测试集F对生成器网络G进行评估,保留最优生成器网络G'和判别器网络D';同时在参数更新结束阶段进行判断,判断训练迭代次数是否已达到最大迭代次数,若已经达到最大迭代次数,则训练阶段结束,进入下一步骤;反之,将进入步骤S3进行循环迭代训练;Step S7: Use the retina test set F to evaluate the generator network G, retain the optimal generator network G' and the discriminator network D'; at the same time, judge at the end stage of parameter update to judge whether the number of training iterations has reached the maximum number of iterations , if the maximum number of iterations has been reached, the training phase ends, and the next step is entered; otherwise, it will enter step S3 for cyclic iteration training; 步骤S8:将视网膜图像样本集F1={ai|i∈[1,RF]}输入最优生成器G'生成血管分割图像,式中RF表示视网膜图像样本集的样本数量;Step S8: Input the retinal image sample set F 1 ={a i |i∈[1,RF ]} into the optimal generator G' to generate a blood vessel segmentation image, where RF represents the number of samples in the retinal image sample set; 所述步骤S2中生成器、判别器、Adam优化器构建并初始化具体为:In the step S2, the generator, the discriminator, and the Adam optimizer are constructed and initialized as follows: 生成器网络G中,网络将扩张路径中得到特征图与收缩路径中得到的高分辨率的特征图相拼接,再对得到的拼接特征图进行卷积操作提取特征,这种方式能够更加充分地利用图像的浅层特征和深层特征;在生成器网络G中也使用了残差模块结构,此结构用来解决深层网络中易出现的退化问题,能够帮助网络更好地学习到图像的特征;为了增大生成器网络G的感受野,本发明在生成器网络G中加入了空洞卷积操作,在不增加网络参数数量的同时增大生成器网络G的感受野;生成器网络G中的卷积核大小都为3x3,在每次下采样操作后,将输出的特征图输入2个残差块ResBlock进行跳跃连接操作,经过4次下采样后得到的特征图使用空洞率分为别为5和3的空洞卷积操作来增大网络的感受野;在上采样的过程中,会拼接浅层网络提取的特征进行操作,网络最后再使用1x1的卷积操作,保证网络输出的颜色通道数和输入颜色通道数相同;In the generator network G, the network splices the feature maps obtained in the expansion path with the high-resolution feature maps obtained in the shrinking path, and then performs a convolution operation on the obtained spliced feature maps to extract features. Use the shallow features and deep features of the image; the residual module structure is also used in the generator network G, this structure is used to solve the degradation problem that is easy to occur in the deep network, and can help the network to better learn the features of the image; In order to increase the receptive field of the generator network G, the present invention adds a hole convolution operation to the generator network G, which increases the receptive field of the generator network G without increasing the number of network parameters; The size of the convolution kernel is 3x3. After each downsampling operation, the output feature map is input into 2 residual blocks ResBlock for skip connection operation. The feature map obtained after 4 times of downsampling is divided into The hole convolution operation of 5 and 3 increases the receptive field of the network; in the process of upsampling, the features extracted by the shallow network will be spliced for operation, and the network will finally use a 1x1 convolution operation to ensure the color channel output by the network. The number is the same as the number of input color channels; 判别器网络D构建具体为:判别器网络D是一个深层卷积神经网络,其主要任务是判断输入的图像是真实图像还是生成器网络G生成的图像;为了防止网络退化问题,本发明在判别器网络D中加入了ResBlock结构,该跳跃连接结构在增加网络深度的情况下依然能够使得网络梯度传播,保证收敛;判别器网络D中每一层卷积核大小为3x3,总共进行4次MaxPooling操作对网络进行下采样,网络的最后使用了一个全连接层进行最终维度的改变,使得最终输出一个[0,1]的置信度q,判断输入的图像是基准的视网膜图像对(a,b),还是生成器网络G生成的视网膜图像对(a,z);The construction of the discriminator network D is as follows: the discriminator network D is a deep convolutional neural network, and its main task is to judge whether the input image is a real image or an image generated by the generator network G; in order to prevent the problem of network degradation, the present invention is used to discriminate. The ResBlock structure is added to the discriminator network D. The skip connection structure can still make the network gradient propagate and ensure convergence when the network depth is increased. The size of each layer of the convolution kernel in the discriminator network D is 3x3, and a total of 4 times MaxPooling is performed The operation downsamples the network. At the end of the network, a fully connected layer is used to change the final dimension, so that a confidence q of [0, 1] is finally output, and the input image is judged to be the benchmark retinal image pair (a, b ), or the retinal image pair (a, z) generated by the generator network G; Adam优化器构建具体为:在Pytorch框架中采用Adam优化方法对训练超参数动态调整,优化训练;在训练过程中,利用Adam优化器智能调整学习率,加快网络的收敛;初始的学习率为0.0002,第一个力矩系数β1为0.5,第二个力矩系数β2为0.999;The Adam optimizer is constructed as follows: the Adam optimization method is used in the Pytorch framework to dynamically adjust the training hyperparameters to optimize the training; during the training process, the Adam optimizer is used to intelligently adjust the learning rate to speed up the convergence of the network; the initial learning rate is 0.0002 , the first moment coefficient β 1 is 0.5, and the second moment coefficient β 2 is 0.999; 所述的步骤S4中生成器网络G生成分割图像具体为:从视网膜训练集E中提取一对视网膜训练样本(ai,bi),i=1,2,3,...,M,将视网膜图像ai输入到生成器网络G当中,对应的图像大小为(3×h×w)其中3表示颜色通道数量,为对应的颜色通道d∈{红,绿,蓝},h表示单张图片的高,w表示单张图片的宽;经过网络的逐层计算,得到生成的视网膜血管分割图zi,图像大小为(1×h×w)其中对应的为单颜色通道的灰度图,即根据图像血管的明显程度,通过图像灰度值大小的形式来表现;In the step S4, the generator network G generates a segmented image specifically as follows: extracting a pair of retinal training samples (a i , b i ) from the retinal training set E, i=1, 2, 3,...,M, The retinal image a i is input into the generator network G, and the corresponding image size is (3×h×w) where 3 represents the number of color channels, which is the corresponding color channel d∈{red, green, blue}, h represents a single The height of a picture, w represents the width of a single picture; through the layer-by-layer calculation of the network, the generated retinal blood vessel segmentation map zi is obtained, and the image size is (1×h×w), which corresponds to the grayscale of a single color channel Figure, that is, according to the apparent degree of blood vessels in the image, it is expressed in the form of the size of the image gray value; 所述的步骤S5中对生成血管分割图像和基准血管分割图像进行分别判别,具体为:判别器网络D同时分别对视网膜训练集E的样本对(ai,bi)和生成的视网膜样本对(ai,zi)进行判别,其中i=1,2,3,...,M,分别给出[0,1]之间的判别置信度q,q表示此样本对为视网膜训练集E中的样本对的概率;再通过损失函数计算生成的血管分割样本zi和基准血管分割样本bi之间的损失值,为下一步反向传播做准备;In the step S5, the generated blood vessel segmentation image and the reference blood vessel segmentation image are respectively discriminated, specifically: the discriminator network D simultaneously separates the sample pair (a i , b i ) of the retinal training set E and the generated retinal sample pair. (a i , z i ) to discriminate, where i=1, 2, 3,...,M, respectively give the discriminant confidence q between [0, 1], q indicates that this sample pair is a retinal training set The probability of the sample pair in E; then calculate the loss value between the generated blood vessel segmentation sample zi and the reference blood vessel segmentation sample b i through the loss function to prepare for the next back propagation; 所述的步骤S6中利用对抗损失对生成器网络G和判别器网络D进行梯度更新具体为:根据给出的损失函数,利用链式求导法则,对生成器网络G和判别器网络D中的参数进行梯度的计算,通过随机梯度下降法,将对应的参数进行更新;对应的参数更新公式为:In the step S6, the gradient update of the generator network G and the discriminator network D using the confrontation loss is specifically: according to the given loss function, using the chain derivation rule, the generator network G and the discriminator network D are updated. The parameters of the gradient are calculated, and the corresponding parameters are updated through the stochastic gradient descent method; the corresponding parameter update formula is:
Figure FDA0003075194480000031
Figure FDA0003075194480000031
其中θt表示生成器网络G和判别器网络D中的第t个组件的参数,η表示超参数中的学习率,
Figure FDA0003075194480000032
表示对应参数的梯度;
where θt denotes the parameters of the t -th component in the generator network G and the discriminator network D, η denotes the learning rate in the hyperparameters,
Figure FDA0003075194480000032
represents the gradient of the corresponding parameter;
所述的步骤S7中利用视网膜图像测试样本对生成器进行评估,保留最优模型参数具体为:在模型本次训练阶段结束时,将视网膜测试集F输入到生成器网络G中,生成器网络G生成视网膜测试合成集F',通过将视网膜测试集F中的样本对(ai,bi)中的基准血管分割图bi和生成的视网膜样本对(ai,zi)中的生成分割图zi进行逐一像素比对,其中i=1,2,3,...,N,每个像素都被分类为血管点和非血管素点;为了对当前生成器网络G进行性能测试,需要通过性能指标客观地进行定量分析;采用精确度(Accuracy,简写Acc)、特异性(Specificity,简写Sp)、敏感度(Sensitivity,简写Se)、Dice系数、F-measure、受试者工作特征曲线(ROC)与坐标轴围成的面积(AUC)和精确率-召回率曲线(PR曲线)与坐标轴围成的面积(mAP)等指标来衡量本文模型的有效性;In the described step S7, the retinal image test sample is used to evaluate the generator, and the optimal model parameters are reserved as follows: at the end of the current training stage of the model, the retinal test set F is input into the generator network G, and the generator network G generates a synthetic retinal test set F' by combining the benchmark blood vessel segmentation map b i in the sample pair (a i , b i ) in the retina test set F with the generated retinal sample pair (a i , z i ) in the generated The segmentation map zi is compared pixel by pixel, where i =1, 2, 3, ..., N, and each pixel is classified as a vessel point and a non-vessel point; in order to test the performance of the current generator network G , which needs to be quantitatively analyzed objectively through performance indicators; accuracy (Accuracy, abbreviated Acc), specificity (Specificity, abbreviated Sp), sensitivity (Sensitivity, abbreviated Se), Dice coefficient, F-measure, receiver work The area enclosed by the characteristic curve (ROC) and the coordinate axis (AUC) and the precision-recall rate curve (PR curve) and the area enclosed by the coordinate axis (mAP) are used to measure the effectiveness of the model in this paper; 其中,AUC多用于医学图像处理的性能测量,AUC的值越接近于1,分割效果就越好;Among them, AUC is mostly used for performance measurement of medical image processing. The closer the value of AUC is to 1, the better the segmentation effect;
Figure FDA0003075194480000041
Figure FDA0003075194480000041
Figure FDA0003075194480000042
Figure FDA0003075194480000042
Figure FDA0003075194480000043
Figure FDA0003075194480000043
Figure FDA0003075194480000044
Figure FDA0003075194480000044
Figure FDA0003075194480000045
Figure FDA0003075194480000045
Figure FDA0003075194480000046
Figure FDA0003075194480000046
其中,TP(true positive)为真阳性,表示被正确分割的血管个数;TN(true negative)为真阴性,表示被正确分割的非血管即背景像素点个数;FP(false positive)为假阳性,即血管被错误地分割成非血管的像素点个数;FN(false negative)为假阴性,即非血管被错误地划分为血管的像素点个数;TP+FN+FP+TN就是图像中感兴趣区域的总像素点个数;Among them, TP (true positive) is a true positive, indicating the number of correctly segmented blood vessels; TN (true negative) is a true negative, indicating the number of correctly segmented non-vessel background pixels; FP (false positive) is false Positive, that is, the number of pixels that blood vessels are wrongly divided into non-vessel; FN (false negative) is false negative, that is, the number of pixels that non-vessel is wrongly divided into blood vessels; TP+FN+FP+TN is the image The total number of pixels in the region of interest in 由于上述性能指标依赖于输出结果的阈值,通过改变真阳分数(Sensitivity)和假阳分数(Specificity)可以绘制ROC数据,而AUC是ROC曲线下方的面积;所有的评价指标都是都考虑在mask内的所有像素上的测试,mask表示视网膜视盘区域;Since the above performance indicators depend on the threshold of the output results, ROC data can be drawn by changing the true positive score (Sensitivity) and the false positive score (Specificity), and AUC is the area under the ROC curve; all evaluation indicators are considered in the mask Test on all pixels within, mask represents the retinal optic disc area; 经过如上指标进行筛选,挑选Acc、Se、Sp、Precision、Recall、F-measure指标皆为最大的模型作为最优模型;最后,在参数更新结束阶段进行判断,判断训练迭代次数是否已达到最大迭代次数,若已经达到最大迭代次数,则训练阶段结束,得到最优生成器网络G'和最优判别器网络D',并进入下一步骤;反之,将进入步骤S3进行继续循环迭代训练;After the above indicators are screened, the model with the largest Acc, Se, Sp, Precision, Recall, and F-measure is selected as the optimal model; finally, it is judged at the end of the parameter update stage to determine whether the number of training iterations has reached the maximum iteration. If the maximum number of iterations has been reached, the training phase ends, the optimal generator network G' and the optimal discriminator network D' are obtained, and the next step is entered; otherwise, step S3 will be entered to continue the loop iteration training; 所述步骤S8中具体为:将视网膜图像样本集F1={ai|i∈[1,RF]},将其中的每一张视网膜图像ai作为最优的生成器网络G中的输入,输出对应的血管分割图像zi,i=1,2,3,...,RF,式中RF表示视网膜图像样本集的样本数量。The step S8 is as follows: taking the retinal image sample set F 1 ={a i |i∈[1,RF ]}, and taking each retinal image a i as the optimal generator network G. Input and output the corresponding blood vessel segmentation image z i , i=1, 2, 3, . . . , RF , where RF represents the number of samples in the retinal image sample set.
CN201910884346.5A 2019-09-19 2019-09-19 An intelligent segmentation method of retinal blood vessel images based on GAN Active CN110807762B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910884346.5A CN110807762B (en) 2019-09-19 2019-09-19 An intelligent segmentation method of retinal blood vessel images based on GAN

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910884346.5A CN110807762B (en) 2019-09-19 2019-09-19 An intelligent segmentation method of retinal blood vessel images based on GAN

Publications (2)

Publication Number Publication Date
CN110807762A CN110807762A (en) 2020-02-18
CN110807762B true CN110807762B (en) 2021-07-06

Family

ID=69487697

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910884346.5A Active CN110807762B (en) 2019-09-19 2019-09-19 An intelligent segmentation method of retinal blood vessel images based on GAN

Country Status (1)

Country Link
CN (1) CN110807762B (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111382785B (en) * 2020-03-04 2023-09-01 武汉精立电子技术有限公司 GAN network model and method for realizing automatic cleaning and auxiliary marking of samples
CN111524142B (en) * 2020-03-10 2023-06-30 浙江工业大学 A Method for Automatic Segmentation of Cerebrovascular Images
WO2021179205A1 (en) * 2020-03-11 2021-09-16 深圳先进技术研究院 Medical image segmentation method, medical image segmentation apparatus and terminal device
CN111429464B (en) * 2020-03-11 2023-04-25 深圳先进技术研究院 Medical image segmentation method, medical image segmentation device and terminal equipment
US11990224B2 (en) * 2020-03-26 2024-05-21 The Regents Of The University Of California Synthetically generating medical images using deep convolutional generative adversarial networks
CN111724344A (en) * 2020-05-18 2020-09-29 天津大学 A method for generating medical ultrasound image data based on adversarial network
CN111833348B (en) * 2020-08-10 2023-07-14 上海工程技术大学 An automatic detection method of vascular deposits based on image processing
CN115035139A (en) * 2021-03-03 2022-09-09 复旦大学 A Sublingual Vein Segmentation Method Based on Generative Adversarial Networks
US11580646B2 (en) 2021-03-26 2023-02-14 Nanjing University Of Posts And Telecommunications Medical image segmentation method based on U-Net
CN114240951B (en) * 2021-12-13 2023-04-07 电子科技大学 Black box attack method of medical image segmentation neural network based on query
CN114648724B (en) * 2022-05-18 2022-08-12 成都航空职业技术学院 Lightweight efficient target segmentation and counting method based on generation countermeasure network
CN116958175B (en) * 2023-09-21 2023-12-26 无锡学院 A method for constructing a blood cell segmentation network and a blood cell segmentation method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109035142A (en) * 2018-07-16 2018-12-18 西安交通大学 A kind of satellite image ultra-resolution method fighting network integration Aerial Images priori
CN109087243A (en) * 2018-06-29 2018-12-25 中山大学 A kind of video super-resolution generation method generating confrontation network based on depth convolution
CN109191471A (en) * 2018-08-28 2019-01-11 杭州电子科技大学 Based on the pancreatic cell image partition method for improving U-Net network
CN109242839A (en) * 2018-08-29 2019-01-18 上海市肺科医院 A kind of good pernicious classification method of CT images Lung neoplasm based on new neural network model
CN110189253A (en) * 2019-04-16 2019-08-30 浙江工业大学 An Image Super-Resolution Reconstruction Method Based on Improved Generative Adversarial Networks

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11257259B2 (en) * 2017-08-15 2022-02-22 Siemens Healthcare Gmbh Topogram prediction from surface data in medical imaging
CN109671094B (en) * 2018-11-09 2023-04-18 杭州电子科技大学 Fundus image blood vessel segmentation method based on frequency domain classification
CN110009576B (en) * 2019-02-28 2023-04-18 西北大学 Mural image restoration model establishing and restoration method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109087243A (en) * 2018-06-29 2018-12-25 中山大学 A kind of video super-resolution generation method generating confrontation network based on depth convolution
CN109035142A (en) * 2018-07-16 2018-12-18 西安交通大学 A kind of satellite image ultra-resolution method fighting network integration Aerial Images priori
CN109191471A (en) * 2018-08-28 2019-01-11 杭州电子科技大学 Based on the pancreatic cell image partition method for improving U-Net network
CN109242839A (en) * 2018-08-29 2019-01-18 上海市肺科医院 A kind of good pernicious classification method of CT images Lung neoplasm based on new neural network model
CN110189253A (en) * 2019-04-16 2019-08-30 浙江工业大学 An Image Super-Resolution Reconstruction Method Based on Improved Generative Adversarial Networks

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Towards Accurate Segmentation of Retinal Vessels and the Optic Disc in Fundoscopic Images with Generative Adversarial Networks;Jaemin Son et al.;《Journal of Digital Imaging》;20181005;第500-509页 *

Also Published As

Publication number Publication date
CN110807762A (en) 2020-02-18

Similar Documents

Publication Publication Date Title
CN110807762B (en) An intelligent segmentation method of retinal blood vessel images based on GAN
CN110197493B (en) Fundus image blood vessel segmentation method
Li et al. Automatic detection of diabetic retinopathy in retinal fundus photographs based on deep learning algorithm
Zhou et al. A benchmark for studying diabetic retinopathy: segmentation, grading, and transferability
CN108717869B (en) Diagnosis aid system for diabetic retinal complications based on convolutional neural network
CN111784671A (en) Pathological image lesion area detection method based on multi-scale deep learning
Zhao et al. Sea-net: Squeeze-and-excitation attention net for diabetic retinopathy grading
WO2022205502A1 (en) Image classification model construction method, image classification method, and storage medium
CN112508864A (en) Retinal vessel image segmentation method based on improved UNet +
CN111833334A (en) Fundus image feature processing and analyzing method based on twin network architecture
Gao et al. Diabetic retinopathy classification using an efficient convolutional neural network
Firke et al. Convolutional neural network for diabetic retinopathy detection
Himami et al. Deep learning in image classification using dense networks and residual networks for pathologic myopia detection
Vinta et al. Segmentation and classification of interstitial lung diseases based on hybrid deep learning network model
CN116503593A (en) Retina OCT image hydrops segmentation method based on deep learning
Thanh et al. A real-time classification of glaucoma from retinal fundus images using AI technology
Zhou et al. A multi-class fundus disease classification system based on an adaptive scale discriminator and hybrid loss
Hussain et al. An Ensemble Deep Learning Model for Diabetic Retinopathy Identification
CN118470031A (en) A multi-level full-resolution feature selection network for retinal blood vessel segmentation
González-Gonzalo et al. Improving weakly-supervised lesion localization with iterative saliency map refinement
Harshini et al. Machine learning approach for various eye diseases using modified voting classifier model
CN117314935A (en) Diffusion model-based low-quality fundus image enhancement and segmentation method and system
de La Torre et al. Diabetic retinopathy detection through image analysis using deep convolutional neural networks
Venneti et al. Amdnet: Age-related macular degeneration diagnosis through retinal fundus images using lightweight convolutional neural network
Mathias et al. Optimising Diabetic Retinopathy Classification using EfficientNetB7

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant