[go: up one dir, main page]

CN112669260B - Method and device for detecting optic disc and macula in fundus images based on deep neural network - Google Patents

Method and device for detecting optic disc and macula in fundus images based on deep neural network Download PDF

Info

Publication number
CN112669260B
CN112669260B CN202011427915.2A CN202011427915A CN112669260B CN 112669260 B CN112669260 B CN 112669260B CN 202011427915 A CN202011427915 A CN 202011427915A CN 112669260 B CN112669260 B CN 112669260B
Authority
CN
China
Prior art keywords
optic disc
model
segmentation
module
macula
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011427915.2A
Other languages
Chinese (zh)
Other versions
CN112669260A (en
Inventor
杨杰
郭天骄
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Jiao Tong University
Original Assignee
Shanghai Jiao Tong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Jiao Tong University filed Critical Shanghai Jiao Tong University
Priority to CN202011427915.2A priority Critical patent/CN112669260B/en
Publication of CN112669260A publication Critical patent/CN112669260A/en
Application granted granted Critical
Publication of CN112669260B publication Critical patent/CN112669260B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Eye Examination Apparatus (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a fundus image optic disc macula lutea detection method and a device based on a deep neural network, which comprises the following steps: giving a data set containing a plurality of fundus color photographs, giving corresponding optic disc center position labels or whole optic disc region labels and macular center position labels, and dividing the data set into a training set and a verification set; preprocessing an original fundus color photograph to obtain an enhanced domain image, labeling a macular position of a video disc, and establishing a pseudo label image; establishing a model for positioning the macula lutea of the optic disc and segmenting the optic disc, and performing model training and verification on a training set and a verification set by utilizing an enhanced domain diagram and a pseudo label diagram; and positioning the optic disc and the macula lutea on the verification set by adopting the model constructed in the S13, extracting the region of interest, and segmenting by using the model to obtain the final macula lutea, optic disc positioning and optic disc segmentation results. The invention simultaneously covers two tasks of optic disc and yellow spot detection, is reliable and is easy to realize.

Description

基于深度神经网络的眼底图像视盘黄斑检测方法及装置Method and device for detecting optic disc and macula in fundus images based on deep neural network

技术领域technical field

本发明涉及计算机视觉与图像处理技术领域,特别涉及一种基于深度神经网络的眼底图像视盘黄斑检测方法及装置。The invention relates to the technical field of computer vision and image processing, in particular to a method and device for detecting optic disc and macula in fundus images based on a deep neural network.

背景技术Background technique

视网膜眼底图像分析与处理是当前计算机辅助诊断领域的一个热点问题,即输入一张眼底彩照,计算机输出眼底的器官结构、病灶等信息,从而辅助医生进行诊疗,同时也可以节约医生的人工成本,实现大规模筛查等。而眼底图像中的视盘、黄斑结构可以提供丰富的医疗信息,因此计算机自动检测视盘、黄斑的算法有着十分重要的意义。目前而言,对于视盘、黄斑的检测工作,按照算法中的图像处理方式可分为基于深度学习的方法和基于传统图像处理的方法。The analysis and processing of retinal fundus images is a hot issue in the field of computer-aided diagnosis at present, that is, input a color photo of the fundus, and the computer outputs information such as the organ structure and lesions of the fundus, so as to assist doctors in diagnosis and treatment, and at the same time save the labor costs of doctors. Realize large-scale screening, etc. The optic disc and macular structure in the fundus image can provide a wealth of medical information, so the computer automatic detection algorithm of the optic disc and macula is of great significance. At present, for the detection of optic disc and macula, according to the image processing method in the algorithm, it can be divided into methods based on deep learning and methods based on traditional image processing.

基于传统图像处理的方法一般是根据视盘、黄斑的颜色信息进行检测,程序复杂且泛化性差,已逐步被近些年新兴的基于深度学习的方法所取代。而基于深度学习的方法中,大致可以分为两个思路,一是根据视盘或黄斑区域颜色特征进行检测,二是根据眼底结构信息进行检测。基于颜色特征进行检测的特点是思路简单,在正常成像条件、病变轻微的条件下结果较为精确,但成像条件差或是病变严重时可能失效。基于结构信息进行检测的特点是较为鲁棒,但通常在非极端情况下结果较前者差。The methods based on traditional image processing are generally based on the color information of the optic disc and macula. The procedure is complex and the generalization is poor. It has been gradually replaced by the emerging method based on deep learning in recent years. In the method based on deep learning, it can be roughly divided into two ideas, one is to detect based on the color characteristics of the optic disc or macular area, and the other is to detect based on the structural information of the fundus. The feature of detection based on color features is that the idea is simple, and the results are more accurate under normal imaging conditions and mild lesions, but may fail under poor imaging conditions or severe lesions. The characteristic of detection based on structural information is more robust, but the results are usually worse than the former in non-extreme cases.

经检索,现有技术中存在以下几种:After retrieval, there are the following types in the prior art:

[1]B.Al-Bander,W.Al-Nuaimy,B.M.Williams,and Y.Zheng,“Multiscalesequential convolutional neural networks for simultaneous detection of foveaand optic disc,”Biomedical Signal Processing and Control,vol.40,pp.91–101,2018。[1] B.Al-Bander, W.Al-Nuaimy, B.M.Williams, and Y.Zheng, "Multiscalesequential convolutional neural networks for simultaneous detection of fovea and optic disc," Biomedical Signal Processing and Control, vol.40, pp.91 –101, 2018.

[2]X.Meng,X.Xi,L.Yang,G.Zhang,Y.Yin,and X.Chen,“Fast and effectiveoptic disk localization based on convolutional neural network,”Neurocomputing,vol.312,pp.285–295,2018.[2] X.Meng, X.Xi, L.Yang, G.Zhang, Y.Yin, and X.Chen, "Fast and effective optical disk localization based on convolutional neural network," Neurocomputing, vol.312, pp.285 –295, 2018.

[3]K.-K.Maninis,J.Pont-Tuset,P.Arbelaez,and L.Van Gool,“Deep retinalimage understanding,”in International conference on medical image computingand computer-assisted intervention.Springer,2016,pp.140–148[3] K.-K.Maninis, J.Pont-Tuset, P.Arbelaez, and L.Van Gool, “Deep retinal image understanding,” in International conference on medical image computing and computer-assisted intervention. Springer, 2016, pp. 140–148

[4]J.Son,S.J.Park,and K.-H.Jung,“Towards accurate segmentation ofretinal vessels and the optic disc in fundoscopic images with generativeadversarial networks,”Journal of digital imaging,vol.32,no.3,pp.499–512,2019.[4] J.Son, S.J.Park, and K.-H.Jung, "Towards accurate segmentation of retinal vessels and the optic disc in fundoscopic images with generative adversarial networks," Journal of digital imaging, vol.32, no.3, pp .499–512, 2019.

[5]Z.Gu,J.Cheng,H.Fu,K.Zhou,H.Hao,Y.Zhao,T.Zhang,S.Gao,and J.Liu,“Ce-net:Context encoder network for 2d medical image segmentation,”IEEETransactions on Medical Imaging,pp.1–1,2019.[5] Z.Gu, J.Cheng, H.Fu, K.Zhou, H.Hao, Y.Zhao, T.Zhang, S.Gao, and J.Liu, "Ce-net: Context encoder network for 2d medical image segmentation,” IEEE Transactions on Medical Imaging, pp.1–1, 2019.

在[1]中,作者将视盘与黄斑定位视为回归问题,设计了一个卷积神经网络(CNN),同时预测二者的位置。但此类方法无法通过模型输出的结果判断结果的可靠性。在[2]中,作者引入了异常检测思想,但需要大量的数据集支撑,不适合推广。在[3][4][5]中,各个作者设计了。不同的网络和学习算法以进行视盘区域分割,但实验只针对视盘区域的图像块,没有视盘区域的图像块的定位过程。In [1], the author regards the positioning of the optic disc and macula as a regression problem, and designs a convolutional neural network (CNN) to simultaneously predict the position of both. However, such methods cannot judge the reliability of the results through the output results of the model. In [2], the author introduced the idea of anomaly detection, but it requires a large amount of data set support and is not suitable for promotion. In [3][4][5], the various authors designed . Different networks and learning algorithms are used to segment the optic disc area, but the experiment is only for the image blocks of the optic disc area, and there is no positioning process of the image blocks of the optic disc area.

以上方法都有各自的局限性。因此,急需提供一种同时涵盖视盘、黄斑检测两个任务,既可靠,又易于实现的眼底图像视盘黄斑检测。The above methods have their own limitations. Therefore, there is an urgent need to provide a reliable and easy-to-implement detection of the optic disc and macula in fundus images that simultaneously covers the two tasks of optic disc and macular detection.

发明内容Contents of the invention

本发明针对上述现有技术中存在的问题,提出一种基于深度神经网络的眼底图像视盘黄斑检测方法及装置,同时涵盖了视盘、黄斑检测两个任务,设计了新算法,兼顾颜色信息和结构信息,使得视盘、黄斑检测各项指标得到了提升,同时本发明的实验均在公开数据集上进行验证,易于复现。Aiming at the problems existing in the above-mentioned prior art, the present invention proposes a method and device for detecting the optic disc and macula in fundus images based on a deep neural network, covering two tasks of detecting the optic disc and macula at the same time, and designing a new algorithm that takes into account both color information and structure information, so that the optic disc and macular detection indicators have been improved, and at the same time, the experiments of the present invention are verified on public data sets, which are easy to reproduce.

为解决上述技术问题,本发明是通过如下技术方案实现的:In order to solve the problems of the technologies described above, the present invention is achieved through the following technical solutions:

本发明提供一种基于深度神经网络的眼底图像视盘黄斑检测方法,其包括:The invention provides a method for detecting the optic disc and macula of the fundus image based on a deep neural network, which includes:

S11:建立数据集;S11: establish a data set;

给定包含若干张眼底彩照的数据集,并给定相应的视盘中心位置标注或整个视盘区域标注、黄斑中心位置标注,并将所述数据集划分为训练集以及验证集;A data set containing several color fundus photos is given, and the corresponding optic disc center position label or the entire optic disc area label, macular center position label is given, and the data set is divided into a training set and a verification set;

S12:预处理;S12: preprocessing;

对原始眼底彩照进行预处理得到增强域图,对视盘黄斑位置标注,建立伪标签图;Preprocess the original color fundus photos to obtain the enhanced domain map, mark the position of the optic disc and macula, and establish a pseudo-label map;

S13:建立模型;S13: building a model;

建立用于视盘黄斑定位以及视盘分割的模型,利用所述S12得到的增强域图以及伪标签图在所述S11中划分好的训练集以及验证集上进行模型训练以及验证;Establishing a model for optic disc macula location and optic disc segmentation, using the enhanced domain map and pseudo-label map obtained in S12 to perform model training and verification on the training set and verification set divided in S11;

S14:测试模型;S14: test model;

在所述S11中划分好的验证集上采用所述S13构建的模型来定位视盘和黄斑,再提取出感兴趣区域,使用所述S13构建的模型进行分割,得到最终黄斑、视盘定位以及视盘分割结果。On the verification set divided in S11, the model constructed in S13 is used to locate the optic disc and macula, and then the region of interest is extracted, and the model constructed in S13 is used for segmentation to obtain the final macula, optic disc positioning and optic disc segmentation result.

较佳地,所述S12中的预处理包括:高斯滤波以及背景减除。Preferably, the preprocessing in S12 includes: Gaussian filtering and background subtraction.

较佳地,所述S12中的对原始眼底彩照Iori进行预处理得到增强域图Ieh进一步为:Preferably, the enhanced domain map I eh obtained by preprocessing the original fundus color photo I ori in S12 is further:

Ieh=4(Iori-G(σ)*Iori)+0.5,I eh =4(I ori −G(σ)*I ori )+0.5,

其中,G(σ)为高斯滤波器,σ是其方差,*表示图像卷积运算,设定σ为图像视场半径的1/30。Among them, G(σ) is a Gaussian filter, σ is its variance, * represents an image convolution operation, and σ is set to be 1/30 of the radius of the image field of view.

较佳地,所述S12中的建立伪标签图进一步包括:生成半径为预设半径,圆心位于视盘和/或黄斑中心的圆形区域。Preferably, the establishment of the pseudo-label map in S12 further includes: generating a circular area with a preset radius and a center located at the center of the optic disc and/or macula.

较佳地,所述S13中模型训练为使损失函数最小化,其中使用的损失函数为利用真实坐标标签与回归网络回归预测结果的差距构建的回归损失函数;利用伪标签图与伪标签分割结果差距构建的分割损失函数;以及利用视盘区域标注与视盘区域分割结构差距构建的分割损失函数。Preferably, the model training in S13 is to minimize the loss function, wherein the loss function used is a regression loss function constructed using the gap between the real coordinate label and the regression prediction result of the regression network; The segmentation loss function constructed by the gap; and the segmentation loss function constructed by the gap between the optic disc region annotation and the optic disc region segmentation structure.

较佳地,所述S13中利用真实坐标标签进行模型训练得到回归网络,利用伪标签图进行模型训练得到伪标签分割网络;利用真实视盘区域标注进行模型训练得到视盘分割网络;进一步地,Preferably, in said S13, the real coordinate labels are used for model training to obtain a regression network, and the pseudo-label map is used for model training to obtain a pseudo-label segmentation network; the model training is performed using real optic disc region labels to obtain an optic disc segmentation network; further,

所述S14中的使用所述S13构建的模型进行测试,包括伪标签分割网络输出的伪标签分割结果、回归网络输出的黄斑、视盘定位结果以及视盘分割结果;Using the model constructed in S13 in S14 to test, including the pseudo-label segmentation result output by the pseudo-label segmentation network, the macula output by the regression network, the optic disc positioning result, and the optic disc segmentation result;

当所述伪标签分割结果的区域形状为规则的类圆形,则将伪标签分割网络输出的伪标签分割结果作为最终的黄斑、视盘定位结果;当所述伪标签分割结构的区域形状为不规则形状,则将回归网络输出的黄斑、视盘定位作为最终的黄斑、视盘定位结果。When the regional shape of the pseudo-label segmentation result is a regular circular shape, the pseudo-label segmentation result output by the pseudo-label segmentation network is used as the final macula and optic disc positioning result; when the regional shape of the pseudo-label segmentation structure is not If the shape is regular, the macula and optic disc positioning output by the regression network will be used as the final macula and optic disc positioning results.

较佳地,所述伪标签分割结果的区域形状根据形状因子SI来判断,所述形状因子SI为:Preferably, the region shape of the pseudo-label segmentation result is judged according to the shape factor SI, and the shape factor SI is:

Figure BDA0002819550810000031
Figure BDA0002819550810000031

其中,C为伪标签分割结果区域的周长,S为伪标签分割结果区域的面积;Among them, C is the perimeter of the pseudo-label segmentation result area, and S is the area of the pseudo-label segmentation result area;

当SI位于预设范围内时,判断其为类圆形,当SI超出预设范围时,判断其为不规则形状。When the SI is within the preset range, it is judged to be a circle-like shape, and when the SI exceeds the preset range, it is judged to be an irregular shape.

较佳地,所述回归损失函数最小化问题定义为:Preferably, the regression loss function minimization problem is defined as:

Figure BDA0002819550810000041
Figure BDA0002819550810000041

其中,θ代表模型参数,n为每次训练的图像数即batch数量。以下假设P表示坐标向量,I表示输出图,下标中pre表示模型预测值,gt表示真实标注值。Among them, θ represents the model parameters, and n is the number of images for each training, that is, the number of batches. The following assumes that P represents the coordinate vector, I represents the output image, pre in the subscript represents the predicted value of the model, and gt represents the real label value.

较佳地,所述分割损失函数最小化问题定义为:Preferably, the segmentation loss function minimization problem is defined as:

Figure BDA0002819550810000042
Figure BDA0002819550810000042

其中,θ代表模型参数,n为每次训练的图像数即batch数量。其余二项损失函数分别为交叉熵损失LCE和Dice系数损失LDiceAmong them, θ represents the model parameters, and n is the number of images for each training, that is, the number of batches. The remaining binomial loss functions are the cross-entropy loss L CE and the Dice coefficient loss L Dice respectively;

交叉熵损失为:The cross entropy loss is:

Figure BDA0002819550810000043
Figure BDA0002819550810000043

Dice系数损失为:The Dice coefficient loss is:

Figure BDA0002819550810000044
Figure BDA0002819550810000044

其中H和W分别表示图像的高和宽,单位为像素,K为类别数,为目标类别数加背景数,坐标(i,j,k)表示图像第i行第j列第k个通道的值。Among them, H and W represent the height and width of the image respectively, the unit is pixel, K is the number of categories, which is the number of target categories plus the number of backgrounds, and the coordinates (i, j, k) represent the k-th channel of the i-th row, j-th column of the image value.

本发明还提供一种基于深度神经网络的眼底图像视盘黄斑检测装置,其用于实现上述所述的基于深度神经网络的眼底图像视盘黄斑检测方法,其包括:数据集建立模块、预处理模块、模型建立模块以及模型测试模块;其中,The present invention also provides a deep neural network-based optic disc macular detection device, which is used to implement the above-mentioned deep neural network-based optic disc macular detection method, which includes: a data set establishment module, a preprocessing module, A model building module and a model testing module; wherein,

所述数据集建立模块用于给定一个包含若干张眼底彩照的数据集,并给定相应的视盘中心位置标注或整个视盘区域标注、黄斑中心位置标注,并将所述数据集划分为训练集以及验证集;The data set building module is used to give a data set containing several fundus color photos, and give the corresponding optic disc center position label or the entire optic disc area label, macular center position label, and divide the data set into a training set and the validation set;

所述预处理模块用于对原始眼底彩照进行预处理得到增强域图,对视盘黄斑位置标注,建立伪标签图;The preprocessing module is used to preprocess the original color fundus photo to obtain an enhanced domain map, mark the position of the macula on the optic disc, and establish a pseudo-label map;

所述模型建立模块用于建立用于视盘黄斑定位以及视盘分割的模型,利用所述预处理模块得到的增强域图以及伪标签图在所述数据集建立模块中划分好的训练集以及验证集上进行模型训练以及验证;The model building module is used to build a model for optic disc macula location and optic disc segmentation, using the enhanced domain map and pseudo-label map obtained by the preprocessing module to divide the training set and verification set in the data set building module model training and validation on

所述模型测试模块用于在所述数据集建立模块中划分好的验证集上采用所述模型建立模块构建的模型来定位视盘和黄斑,再提取出感兴趣区域,使用所述模型建立模块构建的模型进行分割,得到最终的黄斑、视盘定位以及视盘分割结果。The model testing module is used to locate the optic disc and macula using the model built by the model building module on the verification set divided in the data set building module, and then extract the region of interest, and use the model building module to construct The model is segmented to obtain the final results of macula, optic disc positioning and optic disc segmentation.

相较于现有技术,本发明具有以下优点:Compared with the prior art, the present invention has the following advantages:

(1)本发明提供的基于深度神经网络的眼底图像视盘黄斑检测方法及装置,通过对眼底彩照进行标注以及预处理,既考虑了颜色信息,又考虑了结构信息,使定位结果更具有精确性、鲁棒性;(1) The method and device for detecting optic disc and macula in fundus images based on deep neural network provided by the present invention, through labeling and preprocessing color fundus photos, not only color information but also structural information are taken into account, so that the positioning results are more accurate , robustness;

(2)本发明提供的基于深度神经网络的眼底图像视盘黄斑检测方法及装置,通过增强域图以及伪标签图进行模型训练,得到两种网络,提升了分割精度;(2) The fundus image optic disc macula detection method and device based on the deep neural network provided by the present invention carry out model training through the enhanced domain map and the pseudo-label map to obtain two kinds of networks, which improves the segmentation accuracy;

(3)本发明提供的基于深度神经网络的眼底图像视盘黄斑检测方法及装置,同时涵盖了视盘、黄斑两个检测任务,同时实现了黄斑定位以及视盘分割。(3) The deep neural network-based method and device for detecting optic disc and macula in fundus images provided by the present invention simultaneously cover two detection tasks of optic disc and macula, and realize macular positioning and optic disc segmentation at the same time.

当然,实施本发明的任一产品并不一定需要同时达到以上所述的所有优点。Of course, any product implementing the present invention does not necessarily need to achieve all the above-mentioned advantages at the same time.

附图说明Description of drawings

下面结合附图对本发明的实施方式作进一步说明:Embodiments of the present invention will be further described below in conjunction with accompanying drawings:

图1为本发明一实施例的基于深度神经网络的眼底图像视盘黄斑检测方法的流程图;Fig. 1 is the flow chart of the fundus image optic disc macula detection method based on deep neural network of an embodiment of the present invention;

图2为本发明一较佳实施例的伪标签分割网络(NetF)以及视盘分割网络(NetS)的示意图;Fig. 2 is a schematic diagram of a pseudo-label segmentation network (NetF) and a visual disc segmentation network (NetS) of a preferred embodiment of the present invention;

图3为本发明一较佳实施例的输入模块示意图;Fig. 3 is a schematic diagram of an input module of a preferred embodiment of the present invention;

图4为本发明一较佳实施例的ResNet基本模块A示意图;Fig. 4 is a schematic diagram of ResNet basic module A of a preferred embodiment of the present invention;

图5为本发明一较佳实施例的ResNet基本模块B示意图;Fig. 5 is a schematic diagram of ResNet basic module B of a preferred embodiment of the present invention;

图6为本发明一较佳实施例的DAC模块示意图;Fig. 6 is a schematic diagram of a DAC module of a preferred embodiment of the present invention;

图7为本发明一较佳实施例的RMP模块示意图;Fig. 7 is a schematic diagram of the RMP module of a preferred embodiment of the present invention;

图8为本发明一较佳实施例的Dec模块示意图;Fig. 8 is a schematic diagram of a Dec module in a preferred embodiment of the present invention;

图9为本发明一较佳实施例的输出模块示意图;Fig. 9 is a schematic diagram of an output module of a preferred embodiment of the present invention;

图10为本发明一较佳实施例的全连接模块示意图。Fig. 10 is a schematic diagram of a fully connected module in a preferred embodiment of the present invention.

具体实施方式Detailed ways

下面对本发明的实施例作详细说明,本实施例在以本发明技术方案为前提下进行实施,给出了详细的实施方式和具体的操作过程,但本发明的保护范围不限于下述的实施例。The embodiments of the present invention are described in detail below. This embodiment is implemented on the premise of the technical solution of the present invention, and detailed implementation methods and specific operating procedures are provided, but the protection scope of the present invention is not limited to the following implementation example.

如图1所示为本发明一实施例的基于深度神经网络的眼底图像视盘黄斑检测方法的流程图。FIG. 1 is a flow chart of a method for detecting optic disc and macula in a fundus image based on a deep neural network according to an embodiment of the present invention.

请参考图1,本实施例的基于深度神经网络的眼底图像视盘黄斑检测方法包括:Please refer to Fig. 1, the fundus image optic disc macula detection method based on deep neural network of the present embodiment comprises:

S11:建立数据集;S11: establish a data set;

给定一个包含若干张眼底彩照的数据集,并给定相应的视盘中心位置标注或整个视盘区域标注、黄斑中心位置标注,并将数据集划分为训练集以及验证集;Given a data set containing several fundus color photos, and given the corresponding optic disc center position label or the entire optic disc area label, macular center position label, and divide the data set into a training set and a verification set;

S12:预处理;S12: preprocessing;

对原始眼底彩照进行预处理得到增强域图,对视盘黄斑位置标注,建立伪标签图;Preprocess the original color fundus photos to obtain the enhanced domain map, mark the position of the optic disc and macula, and establish a pseudo-label map;

S13:建立模型;S13: building a model;

建立用于视盘黄斑定位以及视盘分割的模型,利用S12得到的增强域图以及伪标签图在S11中划分好的训练集以及验证集上进行模型训练以及验证;Establish a model for optic disc macular location and optic disc segmentation, use the enhanced domain map and pseudo-label map obtained in S12 to perform model training and verification on the training set and verification set divided in S11;

S14:测试模型;S14: test model;

在S11中划分好的验证集上采用S13构建的模型来定位视盘和黄斑,再提取出感兴趣区域,使用S13构建的模型进行分割,得到最终的黄斑、视盘定位以及视盘分割结果。On the verification set divided in S11, the model constructed in S13 was used to locate the optic disc and macula, and then the region of interest was extracted, and the model constructed in S13 was used for segmentation to obtain the final results of macula, optic disc positioning and optic disc segmentation.

较佳实施例中,数据集的黄斑的位置标定,即坐标

Figure BDA0002819550810000061
同时包含视盘位置标定
Figure BDA0002819550810000062
对于将进行视盘分割的数据,需包含完整的视盘区域图Igt。一实施例中,模型训练使用4折交叉验证,即将数据集划分为四份,其中三份作为训练集,另一份作为验证集,重复该过程直到四份数据均作为验证集一次。In a preferred embodiment, the position calibration of the macula of the data set, that is, the coordinates
Figure BDA0002819550810000061
Also includes disc position calibration
Figure BDA0002819550810000062
For the data to be divided into optic discs, the complete optic disc area map I gt should be included. In one embodiment, model training uses 4-fold cross-validation, that is, the data set is divided into four parts, three of which are used as training sets, and the other is used as a verification set, and this process is repeated until all four data sets are used as a verification set once.

较佳实施例中,S12中的预处理包括:高斯滤波、背景减除。S12具体地包括:对原始的RGB眼底彩图Iori,对其进行预处理得到增强图Ieh,该过程可表示为:In a preferred embodiment, the preprocessing in S12 includes: Gaussian filtering and background subtraction. S12 specifically includes: preprocessing the original RGB fundus image I ori to obtain an enhanced image I eh , the process can be expressed as:

Ieh=4(Iori-G(σ)*Iori)+0.5I eh =4(I ori -G(σ)*I ori )+0.5

其中G(σ)为高斯滤波器,σ是其方差,*表示图像卷积运算,设定σ为图像视场半径的1/30。Among them, G(σ) is a Gaussian filter, σ is its variance, * indicates image convolution operation, and σ is set to 1/30 of the radius of the image field of view.

较佳实施例中,S12中的建立伪标签图具体包括:对视盘和黄斑的位置标定

Figure BDA0002819550810000063
Figure BDA0002819550810000064
生成一张视盘的伪标签图
Figure BDA0002819550810000065
Figure BDA0002819550810000066
为包含一个圆心位于
Figure BDA0002819550810000067
半径为
Figure BDA0002819550810000068
Figure BDA0002819550810000069
距离的五分之一的圆的二值图。同理可生成一张黄斑的伪标签图
Figure BDA00028195508100000610
In a preferred embodiment, the establishment of the pseudo-label map in S12 specifically includes: calibrating the positions of the optic disc and macula
Figure BDA0002819550810000063
and
Figure BDA0002819550810000064
Generate a pseudo-label map of the optic disc
Figure BDA0002819550810000065
Figure BDA0002819550810000066
To contain a circle centered at
Figure BDA0002819550810000067
Radius is
Figure BDA0002819550810000068
arrive
Figure BDA0002819550810000069
Binary map of circles of one-fifth of the distance. In the same way, a pseudo-label map of the macula can be generated
Figure BDA00028195508100000610

较佳实施例中,S13中的模型训练包括:利用真实坐标标签进行模型训练得到回归网络(NetP),利用伪标签图进行模型训练得到伪标签分割网络(NetF),利用真实视盘区域标注进行模型训练得到视盘分割网络(NetS)。模型训练包括:用于特征提取的输入模块、Res模块,用于编码的DAC模块、RMP模块,和用于解码的Dec模块、输出模块和全连接模块。将输入模块、Res模块、DAC模块、RMP模块、Dec模块、输出模块连接构成伪标签分割网络(NetF)和视盘分割网络(NetS),利用伪标签或视盘区域与伪标签或视盘分割结果差距构建损失函数,输入训练集训练模型,迭代更新网络参数。将输入模块、Res模块、DAC模块、RMP模块、全连接模块连接构成回归网络(NetP),利用真实位置标定与回归结果差距构建损失函数,输入训练集和验证集训练模型,迭代更新网络参数。伪标签分割网络(NetF)和视盘分割网络(NetS)的网络模型示意图如图2所示,其中,Res表示Res模块,DAC表示密集孔洞卷积模块(Dense Atrous Convolution module,DAC),RMP表示残差多核池化模块(Residual Multi-kernel pooling,RMP),Dec表示解码器模块,

Figure BDA0002819550810000071
表示张量加法。In a preferred embodiment, the model training in S13 includes: using the real coordinate labels to carry out model training to obtain a regression network (NetP), using the pseudo-label map to perform model training to obtain a pseudo-label segmentation network (NetF), and using the real visual disc area label to perform model The optic disc segmentation network (NetS) is obtained through training. Model training includes: input module, Res module for feature extraction, DAC module, RMP module for encoding, Dec module, output module and fully connected module for decoding. Connect the input module, Res module, DAC module, RMP module, Dec module, and output module to form a pseudo-label segmentation network (NetF) and an optic disc segmentation network (NetS). Loss function, input the training set to train the model, iteratively update the network parameters. Connect the input module, Res module, DAC module, RMP module, and fully connected module to form a regression network (NetP), use the gap between the real position calibration and the regression result to construct a loss function, input the training set and verification set to train the model, and update the network parameters iteratively. The schematic diagram of the network model of pseudo-label segmentation network (NetF) and disc segmentation network (NetS) is shown in Figure 2, where Res represents the Res module, DAC represents the Dense Atrous Convolution module (DAC), and RMP represents the residual Residual Multi-kernel pooling (RMP), Dec represents the decoder module,
Figure BDA0002819550810000071
Represents tensor addition.

较佳实施例中,S13中的输入模块用于将两路图像进行卷积、批标准化、非线性激励、最大池化与加和处理。具体的是:RGB域眼底图像和经过第二步预处理后的增强域图像分别通过上下两路输入至输入模块,输入模块中两条路径处理方式相同,依次经过卷积(Conv)、批标准化(BN)、线性整流函数激励(Relu)和最大池化(MaxPool)。其中卷积层的卷积核大小为7×7,输入通道数为3,输出通道数为64,步长为2,填充像素(pad)为3。经过上述处理,两路分别得到RGB域眼底图像和增强域图像的特征图,将两特征图加和后输出至下一模块。过程如图3所示。In a preferred embodiment, the input module in S13 is used to perform convolution, batch normalization, non-linear excitation, maximum pooling and summing on the two images. Specifically, the fundus image in the RGB domain and the enhanced domain image after the second preprocessing step are respectively input to the input module through the upper and lower channels. (BN), linear rectification function excitation (Relu) and maximum pooling (MaxPool). The convolution kernel size of the convolutional layer is 7×7, the number of input channels is 3, the number of output channels is 64, the step size is 2, and the filling pixel (pad) is 3. After the above processing, the two channels respectively obtain the feature maps of the fundus image in the RGB domain and the enhancement domain image, and output the two feature maps to the next module after summing. The process is shown in Figure 3.

较佳实施例中,S13中的输入模块输出的特征图经过4个Res模块提取更深层次的特征。Res模块为ResNet提出的模块,由一系列ResNet基本模块串联构成。ResNet基本模块可分为A、B两种,分别如图4、图5所示。Res1模块由3个ResNet基本模块B串联构成;Res2模块由1个ResNet基本模块A和3个ResNet基本模块B串联构成;Res3模块由1个ResNet基本模块A和5个ResNet基本模块B串联构成;Res4模块由1个ResNet基本模块A和2个ResNet基本模块B串联构成。图4、图5中Conv3x3、Conv1x1分别表示卷积核大小为3×3、1×1的卷积,s=2表示步长为2,即含有下采样操作。In a preferred embodiment, the feature map output by the input module in S13 passes through 4 Res modules to extract deeper features. The Res module is a module proposed by ResNet, which is composed of a series of ResNet basic modules connected in series. The basic modules of ResNet can be divided into two types, A and B, as shown in Figure 4 and Figure 5 respectively. The Res1 module is composed of 3 ResNet basic modules B in series; the Res2 module is composed of 1 ResNet basic module A and 3 ResNet basic modules B in series; the Res3 module is composed of 1 ResNet basic module A and 5 ResNet basic modules B in series; The Res4 module consists of one ResNet basic module A and two ResNet basic module B connected in series. Conv3x3 and Conv1x1 in Figure 4 and Figure 5 represent convolutions with a convolution kernel size of 3×3 and 1×1, respectively, and s=2 means that the step size is 2, which includes a downsampling operation.

Res1模块中的所有卷积输入通道数为上一级的特征图通道数,即64,输出通道数也均为64;Res2模块中步长为2的两个卷积层输入通道数为64,输出通道数为128,其余卷积层输入、输出通道数均为128;Res3模块中步长为2的两个卷积层输入通道数为128,输出通道数为256,其余卷积层输入、输出通道数均为256;Res4模块中步长为2的两个卷积层输入通道数为256,输出通道数为512,其余卷积层输入、输出通道数均为512。The number of input channels of all convolutions in the Res1 module is the number of feature map channels of the upper level, that is, 64, and the number of output channels is also 64; the number of input channels of the two convolutional layers with a step size of 2 in the Res2 module is 64, The number of output channels is 128, and the number of input and output channels of the other convolutional layers is 128; the number of input channels of the two convolutional layers with a step size of 2 in the Res3 module is 128, the number of output channels is 256, and the input and output channels of the other convolutional layers are The number of output channels is 256; the number of input channels of the two convolutional layers with a step size of 2 in the Res4 module is 256, the number of output channels is 512, and the number of input and output channels of the other convolutional layers is 512.

较佳实施例中,S13中的DAC模块包含一系列不同卷积核大小、孔洞大小的孔洞卷积层。RMP模块包含尺度分别为2、3、5和6的池化,尺度分别为2、3、5和6的上采样,和张量拼接过程。具体的是:Res4模块提取的14×14,512通道的特征图,经过DAC模块提取出不同尺度的信息,得到一个含有不同尺度信息的14×14,512通道的特征图;经RMP模块多尺度池化和上采样,再在通道维度上进行张量拼接,得到一个14×14,516通道的特征图。DAC模块如图6所示,图中Conv前的数字表示卷积核大小,rate表示孔洞大小,channel表示输入和输出通道数。RMP模块如图7所示,图中pooling表示最大池化操作,前面的数字表示池化尺度,Conv前的数字表示卷积核大小,卷积输入通道为512,输出通道为1,Upsample表示上采样,Concatenate表示张量拼接。In a preferred embodiment, the DAC module in S13 includes a series of atrous convolution layers with different convolution kernel sizes and hole sizes. The RMP module includes pooling at scales 2, 3, 5, and 6, upsampling at scales 2, 3, 5, and 6, and tensor concatenation. Specifically: the feature map of 14×14, 512 channels extracted by the Res4 module, the information of different scales is extracted through the DAC module, and a feature map of 14×14, 512 channels containing information of different scales is obtained; multi-scale information is obtained by the RMP module Pooling and upsampling, and tensor splicing in the channel dimension to obtain a 14×14, 516-channel feature map. The DAC module is shown in Figure 6. The number before Conv in the figure indicates the size of the convolution kernel, rate indicates the size of the hole, and channel indicates the number of input and output channels. The RMP module is shown in Figure 7. Pooling in the figure indicates the maximum pooling operation, the number in front indicates the pooling scale, the number in front of Conv indicates the size of the convolution kernel, the convolution input channel is 512, the output channel is 1, and Upsample indicates up Sampling, Concatenate means tensor splicing.

较佳实施例中,S13中的RMP模块输出的特征图将经过4个Dec模块进行解码。Dec模块示意图如图8所示。图中ConvT3x3表示卷积核大小为3×3的反卷积操作,步长为2。具体的是:对于多个Dec模块,设每个Dec模块输入的特征图通道数为Ch,则Dec模块第一个Conv1x1卷积层输入通道数为Ch,输出的通道数为

Figure BDA0002819550810000081
[·]表示向下取整;Dec模块第一个反卷积层输入、输出的通道数均为
Figure BDA0002819550810000082
Dec模块第二个Conv1x1卷积层输入通道数为
Figure BDA0002819550810000083
输出的通道数为固定值,在本实施例中4个Dec模块该值依次为256、128、64、64。前3个Dec模块输出的特征图将分别与Res3、Res2和Res1模块输出的特征图做张量相加后输入下一个Dec模块。如图2所示。In a preferred embodiment, the feature map output by the RMP module in S13 will be decoded by four Dec modules. The schematic diagram of the Dec module is shown in Figure 8. ConvT3x3 in the figure represents a deconvolution operation with a convolution kernel size of 3×3 and a step size of 2. Specifically: for multiple Dec modules, let the number of feature map channels input by each Dec module be Ch, then the number of input channels of the first Conv1x1 convolutional layer of the Dec module is Ch, and the number of output channels is
Figure BDA0002819550810000081
[ ] indicates rounding down; the number of input and output channels of the first deconvolution layer of the Dec module is
Figure BDA0002819550810000082
The number of input channels of the second Conv1x1 convolutional layer of the Dec module is
Figure BDA0002819550810000083
The number of output channels is a fixed value. In this embodiment, the values of the four Dec modules are 256, 128, 64, and 64 in sequence. The feature maps output by the first three Dec modules will be added to the feature maps output by the Res3, Res2 and Res1 modules respectively, and input to the next Dec module. as shown in picture 2.

较佳实施例中,末尾的Dec模块输出的特征图经过输出模块解析得到最终的输出结果,输出模块示意图如图9所示。具体的是:图中ConvT4x4表示卷积核大小为4×4的反卷积层,步长为2,本发明中该层输入通道为64,输出通道为32;Conv3x3表示卷积核大小为3×3的卷积层,本发明中该层输入通道为32,输出通道为32;Conv1x1表示卷积核大小为1×1的卷积层,本发明中该层输入通道为32,输出通道数等于问题的类别数加一,对于伪标签分割网络(NetF),输出通道数为3,对于视盘分割网络(NetS),输出通道数为2。In a preferred embodiment, the feature map output by the Dec module at the end is analyzed by the output module to obtain the final output result. The schematic diagram of the output module is shown in FIG. 9 . Specifically: ConvT4x4 in the figure represents a deconvolution layer with a convolution kernel size of 4×4, with a step size of 2. In the present invention, the input channel of this layer is 64, and the output channel is 32; Conv3x3 represents a convolution kernel size of 3 The convolutional layer of × 3, the input channel of this layer is 32 in the present invention, and the output channel is 32; Conv1x1 represents the convolutional layer whose convolution kernel size is 1 × 1, and the input channel of this layer is 32 in the present invention, and the number of output channels Equal to the number of categories of the problem plus one, the number of output channels is 3 for the pseudo-label segmentation network (NetF), and 2 for the disc segmentation network (NetS).

较佳实施例中,S13中的回归损失函数最小化问题定义为:In a preferred embodiment, the regression loss function minimization problem in S13 is defined as:

Figure BDA0002819550810000084
Figure BDA0002819550810000084

其中,θ代表模型参数,n为每次训练的图像数即batch数量。以下假设P表示坐标向量,I表示输出图,下标中pre表示模型预测值,gt表示真实标注值。Among them, θ represents the model parameters, and n is the number of images for each training, that is, the number of batches. The following assumes that P represents the coordinate vector, I represents the output image, pre in the subscript represents the predicted value of the model, and gt represents the real label value.

较佳实施例中,S13中的分割损失函数最小化问题定义为:In a preferred embodiment, the segmentation loss function minimization problem in S13 is defined as:

Figure BDA0002819550810000085
Figure BDA0002819550810000085

其中,θ代表模型参数,n为每次训练的图像数即batch数量。其余二项损失函数分别为交叉熵损失LCE和Dice系数损失LDiceAmong them, θ represents the model parameters, and n is the number of images for each training, that is, the number of batches. The remaining binomial loss functions are the cross-entropy loss L CE and the Dice coefficient loss L Dice respectively;

交叉熵损失为:The cross entropy loss is:

Figure BDA0002819550810000091
Figure BDA0002819550810000091

Dice系数损失为:The Dice coefficient loss is:

Figure BDA0002819550810000092
Figure BDA0002819550810000092

其中H和W分别表示图像的高和宽,单位为像素,K为类别数,为目标类别数加背景数,坐标(i,j,k)表示图像第i行第j列第k个通道的值。一实施例中H和W均设定为448,对于伪标签分割网络(NetF),K为3,对于视盘分割网络(NetS),K为2。Among them, H and W represent the height and width of the image respectively, the unit is pixel, K is the number of categories, which is the number of target categories plus the number of backgrounds, and the coordinates (i, j, k) represent the k-th channel of the i-th row, j-th column of the image value. In one embodiment, both H and W are set to 448. For the pseudo-label segmentation network (NetF), K is 3, and for the video disc segmentation network (NetS), K is 2.

较佳实施例中,回归网络(NetP)输入模块、Res模块、DAC模块、RMP模块与(NetF)、(NetS)中的相应模块结构和连接方式完全相同。RMP模块后连接全连接模块,全连接模块结构示意图如图10所示。In a preferred embodiment, the regression network (NetP) input module, Res module, DAC module, RMP module have the same structure and connection mode as the corresponding modules in (NetF) and (NetS). The RMP module is connected to the fully connected module, and the schematic diagram of the fully connected module is shown in Figure 10.

较佳实施例中,S14中的使用S13构建的模型进行分割包括伪标签分割网络输出的伪标签分割结果以及回归网络输出的黄斑、视盘定位以及视盘分割结果。当所述伪标签分割结果的区域形状为规则的类圆形,则将伪标签分割网络(NetF)输出的伪标签分割结果作为最终的黄斑、视盘定位结果;当所述伪标签分割结构的区域形状为不规则形状,则将回归网络(NetP)输出的黄斑、视盘定位作为最终的黄斑、视盘定位结果。In a preferred embodiment, the segmentation in S14 using the model constructed in S13 includes the pseudo-label segmentation results output by the pseudo-label segmentation network and the macula, optic disc location, and optic disc segmentation results output by the regression network. When the regional shape of the pseudo-label segmentation result is a regular circular shape, the pseudo-label segmentation result output by the pseudo-label segmentation network (NetF) is used as the final macula and optic disc positioning result; when the area of the pseudo-label segmentation structure If the shape is irregular, the macular and optic disc localization output by the regression network (NetP) will be used as the final macular and optic disc localization results.

进一步地,伪标签分割结果的区域形状根据形状因子SI来判断,形状因子SI为:Further, the region shape of the pseudo-label segmentation result is judged according to the shape factor SI, and the shape factor SI is:

Figure BDA0002819550810000093
Figure BDA0002819550810000093

其中,C为伪标签分割结果区域的周长,S为伪标签分割结果区域的面积;Among them, C is the perimeter of the pseudo-label segmentation result area, and S is the area of the pseudo-label segmentation result area;

当SI位于预设范围内时,判断其为类圆形,当SI超出预设范围时,判断其为不规则形状。理论上,若伪标签分割结果区域为圆形,则SI为4π,SI过高或过低均能说明分割结果区域形状不规则。一实施例中设定两个阈值Tmin=11和Tmax=12.2,当Tmin≤SI≤Tmax时取NetF输出的伪标签分割区域的中心作为视盘、黄斑的定位结果。当SI<Tmin或SI>Tmax,或NetF输出的分割结果包含多个区域时,取NetP输出作为视盘、黄斑的定位结果。When the SI is within the preset range, it is judged to be a circle-like shape, and when the SI exceeds the preset range, it is judged to be an irregular shape. Theoretically, if the pseudo-label segmentation result area is circular, then the SI is 4π. If the SI is too high or too low, it can indicate that the segmentation result area has an irregular shape. In one embodiment, two thresholds T min =11 and T max =12.2 are set. When T min ≤ SI ≤ T max , the center of the pseudo-label segmentation area output by NetF is taken as the positioning result of the optic disc and macula. When SI<T min or SI>T max , or the segmentation result output by NetF contains multiple regions, the output of NetP is taken as the positioning result of optic disc and macula.

较佳实施例中,S14中完成视盘、黄斑定位后,再裁出中心位于视盘中心位置,边长为1.6倍视盘到黄斑距离的图像块,将图像块输入进NetS得到视盘区域的分割结果。In a preferred embodiment, after the optic disc and macula are positioned in S14, an image block whose center is located at the center of the optic disc and whose side length is 1.6 times the distance from the optic disc to the macula is cut out, and the image block is input into NetS to obtain the segmentation result of the optic disc region.

下面结合具体实例对上述实施例的基于深度神经网络的眼底图像视盘黄斑检测方法的效果进行验证。对不同方法进行四折交叉验证后,得到实验结果如表1、表2所示:The effect of the method for detecting the optic disc and macula of the fundus image based on the deep neural network in the above embodiment will be verified below with reference to specific examples. After performing four-fold cross-validation on different methods, the experimental results are shown in Table 1 and Table 2:

表1.不同方法在不同数据验证集上视盘黄斑定位ED对比(最好的结果加粗表示)Table 1. ED comparison of different methods on different data validation sets for optic disc and macula positioning (the best results are indicated in bold)

Figure BDA0002819550810000101
Figure BDA0002819550810000101

表2.不同方法在不同数据验证集上视盘分割Dice对比(最好的结果加粗表示)Table 2. Comparison of different methods for optic disc segmentation Dice on different data validation sets (the best results are indicated in bold)

Figure BDA0002819550810000102
Figure BDA0002819550810000102

从表1和表2的结果可以看出,本发明所提出的视盘黄斑检测方法相比于其他基于深度神经网络的方法,在验证集上的结果更好,与真实标定更接近,而且功能较其他方法更全面。As can be seen from the results in Table 1 and Table 2, compared with other methods based on deep neural networks, the optic disc and macula detection method proposed by the present invention has better results on the verification set, is closer to the real calibration, and has better functions. Other methods are more comprehensive.

一实施例中,还提供一种基于深度神经网络的眼底图像视盘黄斑检测装置,其用于实现上述实施例的基于深度神经网络的眼底图像视盘黄斑检测方法,其包括:数据集建立模块、预处理模块、模型建立模块以及模型测试模块;其中,In one embodiment, a deep neural network-based fundus image optic disc macula detection device is also provided, which is used to implement the deep neural network-based optic disc macula detection method of the above embodiment, which includes: a data set establishment module, a pre-set processing module, model building module and model testing module; wherein,

数据集建立模块用于给定一个包含若干张眼底彩照的数据集,并给定相应的视盘中心位置标注或整个视盘区域标注、黄斑中心位置标注,并将所述数据集划分为训练集以及验证集;The data set building module is used to give a data set containing several fundus color photos, and give the corresponding optic disc center position label or the entire optic disc area label, macular center position label, and divide the data set into a training set and a verification set. set;

预处理模块用于对原始眼底彩照进行预处理得到增强域图,对视盘黄斑位置标注,建立伪标签图;The preprocessing module is used to preprocess the original color fundus photo to obtain an enhanced domain map, mark the position of the optic disc and macula, and establish a pseudo-label map;

模型建立模块用于建立用于视盘黄斑定位以及视盘分割的模型,利用预处理模块得到的增强域图以及伪标签图在数据集建立模块中划分好的训练集以及验证集上进行模型训练以及验证;The model building module is used to build a model for optic disc macula localization and optic disc segmentation, and use the enhanced domain map and pseudo-label map obtained by the preprocessing module to perform model training and verification on the training set and verification set divided in the data set building module ;

模型测试模块用于在所述数据集建立模块中划分好的验证集上采用模型建立模块构建的模型来定位视盘和黄斑,再提取出感兴趣区域,使用模型建立模块构建的模型进行分割,得到最终的黄斑、视盘定位以及视盘分割结果。The model testing module is used to locate the optic disc and macula by using the model built by the model building module on the verification set divided in the data set building module, and then extract the region of interest, and use the model built by the model building module to segment to obtain The final macula, optic disc localization and optic disc segmentation results.

此处公开的仅为本发明的优选实施例,本说明书选取并具体描述这些实施例,是为了更好地解释本发明的原理和实际应用,并不是对本发明的限定。任何本领域技术人员在说明书范围内所做的修改和变化,均应落在本发明所保护的范围内。What is disclosed here are only preferred embodiments of the present invention. The purpose of selecting and describing these embodiments in this description is to better explain the principle and practical application of the present invention, not to limit the present invention. Any modifications and changes made by those skilled in the art within the scope of the description shall fall within the protection scope of the present invention.

Claims (9)

1. A fundus image optic disc macula lutea detection method based on a deep neural network is characterized by comprising the following steps:
s11: establishing a data set;
giving a data set containing a plurality of fundus color photographs, giving corresponding optic disc center position labels or whole optic disc region labels and macular center position labels, and dividing the data set into a training set and a verification set;
s12: pre-treating;
preprocessing an original fundus color photograph to obtain an enhanced domain image, labeling a macular position of a video disc, and establishing a pseudo label image;
s13: establishing a model;
establishing a model for positioning macula lutea of the optic disc and segmenting the optic disc, and performing model training and verification on the training set and the verification set which are divided in the S11 by using the enhanced domain map and the pseudo label map obtained in the S12;
the model in the S13 is trained to minimize a loss function, wherein the loss function is a regression loss function constructed by using the difference between the real coordinate label and the regression network regression prediction result; a segmentation loss function is constructed by utilizing the gap between the pseudo label graph and the pseudo label segmentation result; constructing a segmentation loss function by using the difference between the real optic disc region label and the optic disc region segmentation structure;
in the step S13, model training is carried out by using the real coordinate labels to obtain a regression network, model training is carried out by using the pseudo label graph to obtain a pseudo label segmentation network, and model training is carried out by using the real optic disc region labels to obtain an optic disc segmentation network; respectively inputting a training set and a verification set training model based on the obtained loss function, and iteratively updating each network parameter;
s14: testing the model;
and positioning the optic disc and the macula lutea on the verification set divided in the step S11 by adopting the model constructed in the step S13, extracting the region of interest, and segmenting by using the model constructed in the step S13 to obtain the final macula lutea, optic disc positioning and optic disc segmentation results.
2. The deep neural network-based fundus image optic disc macula detecting method according to claim 1, characterized in that the preprocessing in S12 includes: gaussian filtering and background subtraction.
3. The deep neural network-based fundus image optic disc macula detecting method of claim 2, wherein in the S12 original fundus color photograph I ori Preprocessing to obtain an enhanced domain map I eh Further comprises the following steps:
I eh =4(I ori -G(σ)*I ori )+0.5,
where G (σ) is a gaussian filter, σ is its variance, and σ represents the image convolution operation, and σ is set to 1/30 of the image field radius.
4. The method for detecting the macula of an eye fundus image optic disc based on a deep neural network of claim 1, wherein the establishing of the pseudo tag map in S12 further comprises: and generating a circular area with the radius of a preset radius and the center of the circle positioned in the center of the optic disc and/or the yellow spots.
5. The deep neural network-based fundus image optic disc macula detecting method according to claim 1, characterized in that the model constructed in S13 in S14 is subjected to a test including a pseudo label segmentation result output by a pseudo label segmentation network, macula lutea output by a regression network, a optic disc positioning result, and a optic disc segmentation result;
when the region shape of the pseudo label segmentation result is a regular quasi-circular shape, the pseudo label segmentation result output by the pseudo label segmentation network is used as a final yellow spot and optic disc positioning result; and when the region of the pseudo label segmentation structure is in an irregular shape, positioning the yellow spots and the optic discs output by the regression network to obtain a final yellow spot and optic disc positioning result.
6. The method of detecting a fundus image based on a deep neural network for the macula lutea of an optic disc according to claim 5, wherein the region shape of the pseudo tag segmentation result is determined in accordance with a shape factor SI that is:
Figure FDA0004003896100000021
wherein C is the perimeter of the pseudo label segmentation result region, and S is the area of the pseudo label segmentation result region;
and when the SI is within the preset range, judging that the SI is in a round-like shape, and when the SI exceeds the preset range, judging that the SI is in an irregular shape.
7. The deep neural network-based fundus image optic disc macula detecting method of claim 1, wherein the regression loss function minimization problem is defined as:
Figure FDA0004003896100000022
wherein, theta represents a model parameter, and n is the number of images trained each time, namely the number of batchs; let P denote the coordinate vector, I denote the output graph, pre in the subscript denotes the model prediction value, gt denotes the true annotation value.
8. The method for fundus image-based macular disc detection by a deep neural network according to claim 1, wherein the segmentation loss function minimization problem is defined as:
Figure FDA0004003896100000023
wherein, theta represents a model parameter, and n is the number of images trained each time, namely the number of batchs; the other binomial loss functions are respectively cross entropy loss L CE And Dice coefficient loss L Dice
The cross entropy loss is:
Figure FDA0004003896100000024
the Dice coefficient loss is:
Figure FDA0004003896100000025
wherein H and W respectively represent the height and width of the image, the unit is pixel, K is the number of categories and is the number of target categories plus the number of backgrounds, and the coordinates (i, j, K) represent the value of the ith row and jth column of the image.
9. A fundus image optic disc macula lutea detecting apparatus based on a deep neural network, for realizing the fundus image optic disc macula lutea detecting method based on a deep neural network according to any one of claims 1 to 8, comprising: the system comprises a data set establishing module, a preprocessing module, a model establishing module and a model testing module; wherein,
the data set establishing module is used for giving a data set containing a plurality of fundus color photographs, giving corresponding optic disc center position labels or whole optic disc region labels and macular center position labels, and dividing the data set into a training set and a verification set;
the preprocessing module is used for preprocessing an original fundus color photograph to obtain an enhanced domain image, labeling the position of the macula lutea of a video disc and establishing a pseudo label image;
the model establishing module is used for establishing a model for the macular location of the optic disc and the segmentation of the optic disc, and performing model training and verification on a training set and a verification set which are divided in the data set establishing module by utilizing the enhanced domain map and the pseudo label map which are obtained by the preprocessing module; in the model building module, the model is trained to minimize a loss function, wherein the loss function is a regression loss function built by using the difference between a real coordinate label and a regression network regression prediction result; a segmentation loss function is constructed by utilizing the gap between the pseudo label graph and the pseudo label segmentation result; constructing a segmentation loss function by using the difference between the real optic disc region label and the optic disc region segmentation structure; carrying out model training by using a real coordinate label to obtain a regression network, carrying out model training by using a pseudo label graph to obtain a pseudo label segmentation network, and carrying out model training by using a real optic disc region label to obtain an optic disc segmentation network; respectively inputting a training set and a verification set training model based on the obtained loss function, and iteratively updating each network parameter;
the model testing module is used for positioning the optic disc and the yellow spot on the verification set divided in the data set establishing module by adopting the model established by the model establishing module, extracting the region of interest, and segmenting by using the model established by the model establishing module to obtain the final yellow spot, optic disc positioning and optic disc segmenting results.
CN202011427915.2A 2020-12-07 2020-12-07 Method and device for detecting optic disc and macula in fundus images based on deep neural network Active CN112669260B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011427915.2A CN112669260B (en) 2020-12-07 2020-12-07 Method and device for detecting optic disc and macula in fundus images based on deep neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011427915.2A CN112669260B (en) 2020-12-07 2020-12-07 Method and device for detecting optic disc and macula in fundus images based on deep neural network

Publications (2)

Publication Number Publication Date
CN112669260A CN112669260A (en) 2021-04-16
CN112669260B true CN112669260B (en) 2023-02-28

Family

ID=75401580

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011427915.2A Active CN112669260B (en) 2020-12-07 2020-12-07 Method and device for detecting optic disc and macula in fundus images based on deep neural network

Country Status (1)

Country Link
CN (1) CN112669260B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114821189B (en) * 2022-05-18 2024-06-18 重庆邮电大学 Focus image classification and identification method based on fundus image
CN116777794B (en) * 2023-08-17 2023-11-03 简阳市人民医院 Cornea foreign body image processing method and system

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109598733A (en) * 2017-12-31 2019-04-09 南京航空航天大学 Retinal fundus images dividing method based on the full convolutional neural networks of depth
CN110084824A (en) * 2019-04-26 2019-08-02 山东财经大学 Tongue body image partition method, system, equipment and medium based on symmetrical level set
CN110598652A (en) * 2019-09-18 2019-12-20 上海鹰瞳医疗科技有限公司 Fundus data prediction method and device
CN111028230A (en) * 2019-12-24 2020-04-17 贵州大学 A detection algorithm for optic disc and macular location in fundus images based on YOLO-V3

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109598733A (en) * 2017-12-31 2019-04-09 南京航空航天大学 Retinal fundus images dividing method based on the full convolutional neural networks of depth
CN110084824A (en) * 2019-04-26 2019-08-02 山东财经大学 Tongue body image partition method, system, equipment and medium based on symmetrical level set
CN110598652A (en) * 2019-09-18 2019-12-20 上海鹰瞳医疗科技有限公司 Fundus data prediction method and device
CN111028230A (en) * 2019-12-24 2020-04-17 贵州大学 A detection algorithm for optic disc and macular location in fundus images based on YOLO-V3

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Automatic detection and segmentation of optic disc and fovea in retinal images;Renoh Johnson Chalakkal 等;《IET Image Processing》;20181130;第2100-2110页 *
基于深度学习的眼底图像分割与辅助诊断研究;高宏杰;《中国优秀硕士学位论文全文数据库 信息科技辑》;20200215;正文第3-5章 *
眼底图像中自动定位黄斑中心凹的快速定位方法研究;杨波 等;《中国医学装备》;20190731;第16卷(第7期);第33-36页 *
融合残差注意力机制的UNet视盘分割;侯向丹等;《中国图象图形学报》;20200916(第09期);第1915-1929页 *

Also Published As

Publication number Publication date
CN112669260A (en) 2021-04-16

Similar Documents

Publication Publication Date Title
CN106056595B (en) Based on the pernicious assistant diagnosis system of depth convolutional neural networks automatic identification Benign Thyroid Nodules
CN108510472B (en) Method and device for image processing
WO2018108129A1 (en) Method and apparatus for use in identifying object type, and electronic device
CN114742799B (en) Unknown type defect segmentation method for industrial scenes based on self-supervised heterogeneous network
CN110689525B (en) Method and device for identifying lymph nodes based on neural network
CN111814832B (en) Target detection method, device and storage medium
CN110415212A (en) Abnormal cell detection method, device and computer readable storage medium
TW202013311A (en) Image processing method, electronic equipment and storage medium
CN110020658B (en) Salient object detection method based on multitask deep learning
CN114581646B (en) Text recognition method, device, electronic device and storage medium
CN112669260B (en) Method and device for detecting optic disc and macula in fundus images based on deep neural network
CN112150429A (en) Attention mechanism guided kidney CT image segmentation method
CN114782770A (en) A method and system for license plate detection and license plate recognition based on deep learning
WO2024021461A1 (en) Defect detection method and apparatus, device, and storage medium
CN113095333A (en) Unsupervised feature point detection method and unsupervised feature point detection device
CN110991412A (en) Face recognition method and device, storage medium and electronic equipment
CN116503330A (en) Melanoma skin disease detection method and system based on boundary guided transducer
CN116485806A (en) Segmentation method of hard exudates of retina fundus image
CN117095173A (en) Colorectal cancer H &amp; E staining pathological image semantic segmentation method and system
CN111369494B (en) Method and device for detecting ear density of winter wheat
CN116958221A (en) Cell data analysis method, device, equipment and storage medium
CN116452697A (en) Method and system for detecting secondary pulmonary tuberculosis focus
CN111210398A (en) White blood cell recognition system based on multi-scale pooling
CN113505784A (en) Automatic nail annotation analysis method and device, electronic equipment and storage medium
CN113781475B (en) Method and system for detecting salient human targets in thermal infrared images

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant