Ocular LSTM
Ocular LSTM
Article
Deep-Ocular: Improved Transfer Learning Architecture Using
Self-Attention and Dense Layers for Recognition of
Ocular Diseases
Qaisar Abbas 1, * , Mubarak Albathan 1 , Abdullah Altameem 1 , Riyad Saleh Almakki 1 and Ayyaz Hussain 2
1 College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU),
Riyadh 11432, Saudi Arabia; mmalbathan@imamu.edu.sa (M.A.); altameem@imamu.edu.sa (A.A.);
ralmakki@imamu.edu.sa (R.S.A.)
2 Department of Computer Science, Quaid-i-Azam University, Islamabad 44000, Pakistan;
ayyaz.hussain@qau.edu.pk
* Correspondence: qaabbas@imamu.edu.sa; Tel.: +966-5370-14011
1. Introduction
A key challenge in healthcare is identifying eye disease using fundus images [1]. Any
Copyright: © 2023 by the authors.
ailment or disorder that impairs the eye’s ability to function normally or negatively affects
Licensee MDPI, Basel, Switzerland.
the eye’s visual acuity is referred to as an ocular disease [2]. Almost everyone experiences
This article is an open access article
distributed under the terms and
visual issues at some point in their lives. Some require the care of a specialist, while others
conditions of the Creative Commons
are minor that do not appear on healthcare claims or may be handled at home [3]. Globally,
Attribution (CC BY) license (https:// fundus problems are the main reason why people go blind. The most prevalent ocular
creativecommons.org/licenses/by/ diseases (AMDs) include age-related macular degeneration (AMD), cataracts, glaucoma,
4.0/). and diabetic retinopathy (DR). By 2030, more than 400 million people will have DR, per
Figure 1.
Figure A visual
1. A visual example
example of
of ocular
ocular eye-related
eye-related diseases
diseases studied.
studied.
Diagnostics 2023, 13, 3165 3 of 25
(a) Diabetic retinopathy: The persistently elevated blood sugar levels caused by diabetes
can damage the microscopic blood vessels (capillaries) that transport oxygen and
nutrients to the retina. Diabetic retinopathy affects approximately one-third of people
with diabetes over 50.
(b) Cataract: A cataract is the clouding of the lens of the eye. Cataracts can eventually
contribute to blindness if left untreated. People with diabetes are more likely to
develop cataracts and experience vision loss at a younger age than those without
the condition.
(c) Glaucoma is a group of conditions that can cause an optic nerve injury. The optic nerve
transmits retinal signals for processing to the brain. Glaucoma is frequently (but not
always) caused by increased intraocular pressure. Diabetes substantially increases the
risk of glaucoma in comparison to the general population. Open-angle glaucoma (also
known as “the sneak thief of sight”) and angle-closure glaucoma (which develops
abruptly and is a medical emergency) are the two most common varieties.
Multilabel categorization has been used as a different strategy [19]. Ocular disease
datasets [20,21] could be more balanced. This imbalance makes it difficult to accurately
identify or classify sickness or even a standard retinograph image. This method is not
recommended for broad classification problems due to its low accuracy. The classification
of ocular illnesses is the goal of this effort. It is not suggested to categorize any disease
using the dataset used in this study since it was very unbalanced. This mismatch caused a
lot of variability throughout training, which left a lot to be desired. The strategy we used to
address the issue was balancing the images of the classes. We balanced the ranks by taking
the same number of pictures from each category and feeding them into a pretrained AlexNet
model rather than utilizing all the images and categorizing all the illnesses simultaneously.
As a result, our study initially balanced the dataset by training the classes on the pretrained
AlexNet architecture with the same amount of data for each category. By selecting an equal
number of images for each type, we first loaded the dataset and the associated image into
the dataset. The AlexNet model used the transfer learning (TL) approach in this research.
The accuracy of each class rose once we correctly balanced the dataset.
We developed a deep-ocular detection system based on retinal fundus images. The
advantage of improving the AlexNet (iAlexNet) architecture by incorporating attention
and dense layers, compared to other transfer learning architectures, lies in its ability to
enhance feature extraction, adaptability, and interpretability for specific tasks. To create
this system, we used ODIR and RFMiD datasets, which included various retinographics
of distinct classes of the fundus, using cutting-edge image classification algorithms like
ensemble-based transfer learning. These models, however, only employ one deep learning
network to identify eye-related diseases; hence, the efficacy of this method still has to be
enhanced. In this paper, we suggest a three-step hybrid ensemble model that combines a
classifier, a feature extractor, and a feature selector. The original image features are first
extracted using a pre-trained AlexNet model with an enhanced structure to overcome
the overfitting and gradient vanishing problems. The extracted features are then selected
using the ReliefF method, and then the most crucial elements are chosen to minimize the
feature dimension. Finally, an XgBoost classifier offers classification outcomes based on the
desired features.
tasks where certain regions carry critical information (e.g., diseases in medical images).
Attention mechanisms enable the model to assign varying levels of importance to
different parts of an image, leading to improved feature extraction.
(4) The introduction of additional dense layers further refines the features extracted
by the convolutional layers in the iAlexNet model. Dense layers capture high-level
abstractions and relationships among features, enabling the model to learn more
complex patterns and representations.
(5) The improved iAlexNet architecture allows for the fine-tuning of specific layers, mak-
ing it adaptable to a wide range of tasks. By unfreezing selected layers during training,
the model can specialize its learned representations to match the requirements of
the target task. This adaptability is especially valuable when working with limited
labeled data.
2. Related Work
Ophthalmology has witnessed significant advancements with the advent of artificial
intelligence (AI) and deep learning [22]. This literature review sheds light on pivotal studies
focusing on these advancements, emphasizing the range of diseases tackled, methodologies
employed, and the results achieved. The state-of-the-art studies are described in Table 1.
Table 1. Cont.
In the early days of applying AI to ophthalmology, ref. [23] made strides by proposing
a method for optic disc (OD) localization using convolutional neural networks (CNNs).
Their innovative data preparation and two-stage training process adeptly addressed the
class-imbalance issue, demonstrating a remarkable detection rate of 99.11%. Unique to
this approach was the replacement of the less informative blue channel with segmented
vasculature maps, providing richer context for the CNN structures. Continuing on the trail
of diabetic retinopathy (DR), ref. [24] showcased a hybrid DL network to detect and grade
DR severity. With the utilization of ResNet50, they achieved an accuracy of up to 96% on
the IDRiD dataset. The potential of AI in improving clinical outcomes became even more
evident with [25], where a single network was able to predict the severity of DR with high
sensitivity and specificity.
Expanding the AI spectrum, ref. [26] addressed the global challenge of cataracts, a
predominant cause of blindness. They proposed an efficient cataract detection model
using the VGG19 network, which minimized the need for manual retinal feature extraction.
This research marked a critical transition towards automating ophthalmological diagnosis.
Ref. [27] further highlighted the inherent limitations of manual diagnoses due to constrained
medical resources. They released the comprehensive OIA-ODIR dataset and benchmarked
the performance of nine prominent CNNs, suggesting the importance of feature fusion for
multi-disease classification.
Recognizing the importance of timely examinations, ref. [28] explored the application
of AI in Diabetic Eye Disease (DED). They presented a novel model for multi-class DED
classification, achieving an accuracy of 81.33%. Their work established the potential of
optimized deep learning architectures, especially in scenarios with limited data availabil-
ity. Moving ahead, refs. [29,30] concentrated on diagnosing various eye diseases using
advanced deep learning algorithms, achieving significant accuracies and emphasizing the
efficiency of automated systems.
The global shortage of ophthalmologists is a growing concern, and ref. [31] delineated
the promise of AI in addressing this. The study demonstrated the efficacy of deep learning
in classifying eye diseases even with images of minimal resolution. Following this trend,
ref. [32] detailed a unique approach for detecting glaucoma by employing a CNN based on
the Resnet152 model, highlighting the enhanced accuracy potential of machine learning.
Recent advancements have also underscored the vast database potential, as seen
with [33], in which the authors developed an algorithm using a vast collection of AS-OCT
images. Furthermore, ref. [34] introduced EyeDeep-Net, a multi-layer neural network
designed to efficiently diagnose various eye problems, setting a benchmark in the field. Of
note, ref. [35] provided a nuanced approach, deploying a DCNN optimized with a whale
optimization algorithm. With a significant accuracy leap of 8.1%, this model effectively
detected and classified a range of eye diseases, from cataracts to AMD. Concluding this
review, ref. [36] brought forth a groundbreaking patient-level multi-label ocular disease
baseline
[35] DCNN optimized with whale optimization al- 8.1% increase Used unbalance dataset and model overfitting.
gorithm for the detection of multiple eye dis- in accuracy
eases.
Diagnostics 2023, 13, 3165 6 of 25
[36] DCNet for patient-level multi-label ocular dis- Enhanced Limited eye diseases with huge computational
ease classification with a multi-label soft mar- classification power.
gin loss. performance
classification model, the DCNet. This model, with its three integrated components, pre-
sented a holistic approach to patient diagnostics, further accentuating the role of AI in
3. revolutionizing
Materials andophthalmological
Methods diagnostics.
The overall
3. Materials andsteps
Methodsare represented in Algorithm 1 and a systematic flow diagram of
the proposed system is presented in Figure 2. The deep-ocular system consisted of differ-
The overall steps are represented in Algorithm 1 and a systematic flow diagram of the
entproposed
phases such
systemasispreprocessing, data 2.
presented in Figure augmentation, ansystem
The deep-ocular improved AlexNet
consisted architecture,
of different
feature selection and feature classification. To develop this deep-ocular system,
phases such as preprocessing, data augmentation, an improved AlexNet architecture, fea- we inte-
grated the below
ture selection andsteps,
featureand briefly describe
classification. themthis
To develop in deep-ocular
the subsequent paragraph.
system, we integrated
the below steps, and briefly describe them in the subsequent paragraph.
Input retinographics
Figure 2.2.The
Figure Theproposed
proposed deep-ocular blockdiagram.
deep-ocular block diagram.
The accuracy and efficiency problems in the multiclass diagnosis of glaucoma (GA),
The accuracy and efficiency problems in the multiclass diagnosis of glaucoma (GA),
diabetic retinopathy (DR), cataract (CT), and normal (NL) eye-related diseases were re-
diabetic retinopathy (DR), cataract (CT), and normal (NL) eye-related diseases were re-
solved with the use of an advanced hybrid model containing feature extraction, feature
solved withand
selection, theclassification
use of an advanced
components. hybrid model
This work containing
suggests feature
a three-step extraction,
hybrid ensemble feature
selection, and classification
model including components.
a feature extractor, feature This work
selector, andsuggests
classifier atothree-step
address thishybrid
issue. ensemble
The
model
retinograph image was preprocessed, and features were first extracted using an upgraded issue.
including a feature extractor, feature selector, and classifier to address this
The retinograph
version of iAlexNetimageas awas preprocessed,
feature andReliefF
extractor. The features werethen
method firstassigned
extracted using an up-
a priority
score to
graded each extracted
version feature.
of iAlexNet as The XgBoost
a feature classifier The
extractor. received the first
ReliefF n features
method then as the
assigned a
input after determining the optimal number of input features (n) using the trial-and-error
priority score to each extracted feature. The XgBoost classifier received the first n features
asapproach.
the inputThe XgBoost
after classifier then
determining output the
the optimal classification
number results
of input based on
features theusing
(n) n chosen
the trial-
features, i.e., the DR, CT, GL, and NL classes of eye-related disease.
and-error approach. The XgBoost classifier then output the classification results based on
the n chosen features, i.e., the DR, CT, GL, and NL classes of eye-related disease.
Algorithm 1: Outline for using the improved iAlexNet architecture for feature extraction,
applying the Relief algorithm for feature selection, and then classifying the information features
using XgBoost classifier
Step 1 Let X be the dataset of eye images. Let Y be the corresponding class labels for the
images. AlexNet(), fAlexNet (x) represents the feature extraction process using the
modified AlexNet. Let F be the set of selected features after applying the Relief
algorithm. Let D be the distribution of the data. Let H be the AdaBoost to predict the
retinograph class (DR, CT, GL, and NL).
Step 2 Load pretrained AlexNet model: load a pretrained AlexNet model: mAlexNet .
Diagnostics 2023, 13, 3165 7 of 25
Algorithm 1: Outline for using the improved iAlexNet architecture for feature extraction,
applying the Relief algorithm for feature selection, and then classifying the information features
using XgBoost classifier
Step 3 Improved AlexNet architecture: load AlexNet base model with ImageNet weights,
base_model, incorporate additional convolutional and attention layers, and add dense
layers to refine features: FeaturesEnhancement(Self, FeatureMaps).
Feature extraction: fAlexNet (x) = m − iAlexNet(mAlexNet , x)
Step 4 Fine-tuning and early stopping: unfreeze selected top layers of the model, recompile
model with reduced learning rate: {model.compile(optimizer = Adam(lr = 0.0001),
loss = ‘categorical_crossentropy’, metrics = [‘accuracy’])}.
Implement early stopping to prevent overfitting: {early_stopping =
EarlyStopping(patience = 3, restore_best_weights = True)}.
Step 5 Apply relief algorithm for feature selection: F = Relief(X,Y).
Step 6 Train gradient boosting classifier on selected features:
gb_classifier.fit(X_train_selected, y} {{train}}).
Step 7 Predict class labels using the classifier: {y_pred =
gb_classifier.predict(X_test_selected)}.
Calculate accuracy by comparing predicted and actual labels: {accuracy =
accuracy_score(y}_{{test}},{y_pred)}.
Step 8 Output: display calculated accuracy and predicted class label for the new retinograph.
Figure 3.
Figure 3. A
A visual
visual example
example of
of preprocessing
preprocessing and
and data
data augmentation
augmentation to
to control
control class
class imbalance.
imbalance.
3.3.1. AlexNet
3.3. Model Model for Feature Extraction
Architecture
AlexNet
The improvedis a transfer
AlexNetlearning (TL) algorithm
(iAlexNet) architecturethat can
offers significantly
the above increase computing
advantages. Other
performance using two GPUs for calculations [23]. Since AlexNet
transfer-learning architectures like ResNet, VGG, and Inception can also excel in different is a massive network
with 60 million
scenarios based parameters
on their unique and 650,000 neurons,
characteristics. It itis needs a lot of labeled
recommended examples
to experiment to
with
train [24], which the labeled DR, CAT, GL, and NL image resources
multiple architectures and analyze their performance in the context of the specific task atcannot provide. When
insufficiently
hand. All of the labeled examples
retinograph are available,
images were sent transfer
into AlexNet learning is a practical
for feature and popular
extraction, and the
technique for training deep neural networks. Utilizing every
network’s output from a particular layer was considered for classifying eye-related parameter in a pre-trained
dis-
network as an initialization step might make use of characteristics
eases. Transfer learning allows standard features from trained convolutional neural net- discovered via the use of
enormous datasets. The generated parameters from using these
works to be used as input classifiers for imaging tasks, since most computer-aided diag-layers, primarily utilized
for feature
nosis (CAD) extraction,
systems and can other
aid in medical-image
training convergence. Furthermore,
interpretation systems transfer
cannotlearning may
train convo-
be implemented on regular personal computers, whereas deep network training requires a
lutional neural networks from scratch. The ReliefF method was used to rank the charac-
high-performance GPU and CPU.
teristics that were retrieved in the first phase according to their relevance. The initial fea-
Figure 4 shows the improved AlexNet pretrained model used to recognize DR, CAT,
tures most crucial for categorization were then chosen through trial and error. The
GL and NL ocular diseases. The basic AlexNet is composed of eight layers, categorized
XgBoost model is described in Section 3. The XgBoost model classifies the filtered features
into convolutional, max-pooling, and fully connected layers. The first five layers consist
after receiving the previously selected features as input, producing final classification re-
of convolutional operations, followed by some layers with max-pooling operations. The
sults.
last three layers are fully connected. The network architecture is organized in a way that
involves parallel processing across two GPUs, with the exception of the final layer. To
3.3.1. AlexNet Model for Feature Extraction
develop this improved AlexNet architecture, we added the self-attention and dense layers.
AlexNet
We is a transfer
can represent learning (TL)
the improved algorithm
iAlexNet that canassignificantly
architecture follows: increase compu-
ting performance using two GPUs for calculations [23]. Since AlexNet is a massive net-
# Convolutional layer (1): convolution operation with learnable filters and activa-
work with 60 million parameters and 650,000 neurons, it needs a lot of labeled examples
tion function.
to train [24], which the labeled DR, CAT, GL, and NL image resources cannot provide.
# Max-pooling layer (1): downsampling operation to reduce spatial dimensions.
When insufficiently labeled examples are available, transfer learning is a practical and
# Convolutional Layer (2): another convolution operation with activation.
popular
# technique layer
Max-pooling for training deep neural
(2): additional networks. Utilizing every parameter in a pre-
downsampling.
trained
# network as an
Convolutional initialization
layer step might make
(3): third convolution operationuse of characteristics
with activation. discovered via
the
# use of enormous datasets. The generated parameters
Convolutional layer (4): fourth convolution operation with activation. from using these layers, primar-
ily
# utilized for feature extraction, can aid in training convergence.
Convolutional layer (5): fifth convolution operation with activation. Furthermore, transfer
learning
# may be implemented on regular personal
Max-pooling layer (3): further spatial downsampling. computers, whereas deep network
training
# requires
Flatten layer:aflattening
high-performance
the featureGPU
maps and
to CPU.
a vector for input to fully connected layers.
# Figure 4 shows the
Fully connected improved
layer AlexNet
(1): neurons fullypretrained
connectedmodel to the used to recognize
flattened DR, CAT,
feature vector.
GL
# and NL ocular diseases. The basic AlexNet
Fully connected layer (2): second fully connected layer. is composed of eight layers, categorized
Diagnostics 2023, 13, 3165 10 of 25
# Fully connected layer (3): final fully connected layer (output layer).
# Convolutional layer (6): fifth convolution operation with activation.
into convolutional,
# max-pooling,
One self-attention layer. and fully connected layers. The first five layers consist
of
# convolutional operations, followed
Four convolutional layers (10): fifthbyconvolution
some layersoperation
with max-pooling operations. The
with activation.
last
# three layers are fully
One max-pool layer. connected. The network architecture is organized in a way that
involves
# Oneparallel processing
global average poolacross
layer. two GPUs, with the exception of the final layer. To
develop
# this improved
Two dense layers. AlexNet architecture, we added the self-attention and dense layers.
Figure 4. A modified AlexNet Architecture (iAlexNet) with Relief method and XgBoost classifier to
recognize eye-related
recognize eye-related diseases.
diseases.
k di f f ( F ,R ,H )
W ( Fl ) = W ( Fl ) − ∑ l i i
m.k +
h j = 1 i (1)
P(c)
∑classs( Ri ) 1− p(class( R )) ∑kj=1 di f f ( Fl , Ri , Mi (c)) /Mi
i
classification speed and accuracy, just a small number of the most crucial attributes needed
to be employed. The following sections will provide further examples.
4. Experimental Results
4.1. Experimental Setup
All deep learning (DL) models involved in the study were pre-trained on ImageNet,
and their training and testing were conducted using the publicly available datasets. This
framework was compiled with CUDA 8.0 and CUDNN 5.1 to enhance performance and
efficiency. The experiments were executed on a high-powered workstation with three Intel
CPUs, and a substantial 64 GB of memory. Ensuring smooth and seamless operations, the
workstation operated on the Windows 2022 Server operating system. This robust hardware
and software setup facilitated the comprehensive training and testing of the deep networks,
ensuring accurate and reliable results in the evaluation process.
In evaluating the performance of multi-label eye-disease classification networks, we
employed five assessment metrics, i.e., accuracy (AC), sensitivity (SE), specificity (SP),
the F1-score (F1), and the AUC, and their average, termed as the Final-score. The F1,
representing the harmonic mean of precision and recall, attains high values only when both
precision and recall are elevated. Given that Kappa and F1 focus on a single threshold, and
considering the probabilistic output of classification networks, the AUC (area under the
employed five assessment metrics, i.e., accuracy (AC), sensitivity (SE), specificity (SP), the
F1-score (F1), and the AUC, and their average, termed as the Final-score. The F1, repre-
senting the harmonic mean of precision and recall, attains high values only when both
precision and recall are elevated. Given that Kappa and F1 focus on a single threshold,
Diagnostics 2023, 13, 3165 and considering the probabilistic output of classification networks, the AUC (area 14 of 25under
the ROC curve) was utilized to effectively assess multiple thresholds. These five metrics
were computed using the sklearn package, ensuring a thorough and efficient evaluation
process.
ROC curve) was utilized to effectively assess multiple thresholds. These five metrics were
computed
Python’susing the sklearn
pre-trained package,and
iAlexNet ensuring a thorough
transfer learning and efficient
(TL) evaluation
techniques wereprocess.
used to re-
Python’s pre-trained iAlexNet and transfer learning (TL) techniques were
fine the model. Our proposed system accuracy and loss for both training and validation used to
refine the model. Our proposed system
after 300 epochs are shown in Figure 5. accuracy and loss for both training and validation
after 300 epochs are shown in Figure 5.
Parameters Value
Model initial learning rate 5 × 10−4
Model learning rate and drop factor 0.1
Regularization term: L2 regularization 1 × 10−4
Max number of epochs 100
Mini batch size 32
models are learning from the training dataset. The primary objective of training precision is
to extract the hyperparameters and determine whether our models suffer from overfitting
or underfitting. When we completed training our models with the training dataset and
validated their performance in the validation dataset, only then could we proceed to the test
accuracy, which is the final accuracy of our models. When we refer to accuracy in this paper,
we are referring to test accuracy. Sometimes, veracity alone is not sufficient. We could not
conclude that our model was very accurate based solely on its accuracy because, in this
paper, we had to correctly classify both diseases and no diseases. Those who have a disease
were termed “positive” in terms of deep learning, whereas those who do not have a disease
were termed “negative.” Precision provides a distinct picture of how many disease patients
are correctly identified within the entire dataset. Sometimes, even precision is insufficient.
For instance, if the dataset is highly biased towards one target, recall provides the number
of correctly classified true positives, i.e., those individuals who are truly diseased and for
whom our model made a disease prediction. This recall was the most essential metric for
our research project because if we have a low recall, our model may incorrectly predict a
diseased individual.
The F1 score is often referred to as the harmonic mean of precision and recall. If
someone asserts that precision and recall are of equal importance, then he or she should
focus on the F1 score. In this research, the F1 score was our second-highest priority after
the recall. In addition, confusion matrices were utilized in this study’s model evaluation.
The confusion matrix is a common error matrix used to evaluate the efficacy of supervised
learning algorithms. Each column in a confusion matrix represents the expected category,
and the sum of each column represents the total number of expected data points for that
category. The total quantity of data in each row indicates the number of instances of that
category in the data, and each row displays the data’s actual attribution category.
The accuracy of the classification results may reach 95.33% when the top 20 significant
features are used as the XgBoost classifier’s input, and is the greatest value when compared
to other numbers of input features. Even though accuracy can remain constant when only
the first seven characteristics are entered, a larger number of features results in a longer
training period for the model. The top five characteristics provided by the iAlexNet and
ReliefF algorithms were, therefore, found to be the best ones for this application. Table 5
shows the different TL models to classify ocular diseases by using different epochs and
combination of various architectures.
We conducted numerous independent repeated tests to assess the strength of the
proposed model in terms of accuracy, specificity, and sensitivity compared with several
existing models for the accuracy comparison in the following sections. The best feature
number n was chosen for each separate repeat experiment in accordance with the unique
experimental findings, and the value of n was necessarily set to 300. The model’s generality
will not have been impacted, though, because we only needed to know the value of the
ideal feature number n once throughout the model training process.
Four models were compared in the experimental analysis that follows and is addressed
in Section 4. Improved AlexNet, ReliefF, and XgBoost were used in the suggested model
compared to other state-of-the-art deep learning models. Those models were also con-
structed to test the efficacy of the model components suggested in this paper. In Figure 6,
the model accuracy and loss for both training and validation after 300 epochs are shown.
The experimental findings in Table 6 demonstrate that all models had acceptable accuracy.
The proposed model improved on the original AlexNet’s structure in comparison to the
model. According to the results, the two models’ accuracies were comparable, but the im-
proved model (improved AlexNet) may significantly help the overall model perform better
because if model is built using one model rather than another, it must classify 1000 features,
which significantly lengthens the training time without clearly improving accuracy. A
visual result of the proposed system is presented in Figure 6.
Diagnostics 2023, 13, 3165 16 of 25
Table 6. Classification time consumed by models iAlexNet + XgBoost and iAlexNet + ReliefF +
XgBoost (n = 300).
The proposed model created an XgBoost model to categorize the retrieved features
after using AlexNet to extract the original image’s features. The accuracy of the model was
found to be 95.642 ± 0.398% when the results of the models were compared. The proposed
model also performed better than other architectures in terms of its specificity, sensitivity,
and F-score, demonstrating a significant improvement in performance over the model.
Because the proposed model was an ensemble model that employed XgBoost as a classifier
and AlexNet as a feature extractor, it was superior to the original model. Additionally, the
model we suggested was enhanced based on it. In the model, the AlexNet-extracted features
were further sorted by the ReliefF algorithm, and the trial-and-error method was used to
find the right number of feature inputs to improve the XgBoost classification performance
and, as a result, the model’s overall accuracy. Table 7 also shows that model finished the
classification task faster than the original model did (5.917 ± 0.001 s vs. 5.924 ± 0.001 s),
while the accuracy went up from 95.145 ± 0.404% to 95.642 ± 0.398% (p = 0.05, n = 300).
In terms of model correctness, the proposed model stood out among the models, with an
accuracy rating of 95.642 ± 0.398%.
racy. A visual result of the proposed system is presented in Figure 6.
Table 6. Classification time consumed by models iAlexNet + XgBoost and iAlexNet + ReliefF +
XgBoost (n = 300).
Figure
Figure 6. 6.
AA visualdiagram
visual diagram of
of results
results derived
derivedfrom thethe
from proposed model
proposed for the
model fordetection of fourof four
the detection
classes of eye-related diseases.
classes of eye-related diseases.
The proposed model created an XgBoost model to categorize the retrieved features
Table 7. Classification efficiency of proposed deep-ocular (our proposed model (AlexNet + ReliefF +
after using AlexNet to extract the original image’s features. The accuracy of the model was
XgBoost))
found tomodel compared
be 95.642 ± 0.398%to other
whendifferent architectures
the results in terms
of the models of time in seconds.
were compared. The proposed
Using Different Feature Extractors
Models Total Time/s *
AlexNet + ReliefF + XGBOOST 5.917 ± 0.001
InceptionV3 + ReliefF + XGBOOST 6.428 ± 0.001
SqueezeNet + ReliefF + XGBOOST 5.918 ± 0.002
Using Different Feature Selectors
AlexNet + ReliefF + XGBOOST 5.917 ± 0.001
Models Total time/s
AlexNet + PSO + XGBOOST 5.918 ± 0.001
AlexNet + VGG16 + XGBOOST 5.918 ± 0.001
AlexNet + ReliefF + XGBOOST 5.917 ± 0.001
Using Different Final Classifiers
Models Total time/s
AlexNet + ReliefF + XGBOOST 5.917 ± 0.001
AlexNet + ReliefF + Random Forest 7.271 ± 0.001
AlexNet + ReliefF + ELM 6.029 ± 0.001
* The time required for total classification in seconds.
Diagnostics 2023, 13, 3165 18 of 25
In general, the performance of the four compared models got better with each added
element, proving unequivocally that under the suggested model, each component con-
tributes favorably to the ensemble model’s performance improvement. The integrated
model put out in this study can successfully and precisely carry out the identification and
classification of eye-related diseases.
Table 7 presents the classification efficiency of the proposed deep-ocular model, which
integrated AlexNet, ReliefF, and XgBoost, in comparison to various other architectures
based on the total time taken in seconds for classification. It evaluates the performance
using different feature extractors such as InceptionV3 and SqueezeNet combined with
ReliefF and XgBoost. This table illustrates that the total time taken by the proposed model
(AlexNet + ReliefF + XgBoost) was 5.917 s, with a very minimal standard deviation, and
it was compared with other configurations in which the total times varied. Moreover, the
performance of the proposed model with different feature selectors and final classifiers
was evaluated. The models, including AlexNet combined with other algorithms like PSO,
VGG16, Random Forest, and ELM, were compared in terms of the total classification
Diagnostics 2023, 13, x FOR PEER REVIEW 19 time.
of 26
The proposed model consistently showed a lower total time for classification in comparison
with configurations like AlexNet + ReliefF + Random Forest and AlexNet + ReliefF + ELM,
which took 7.271 s and 6.029 s, respectively, for total classification. The results emphasize
as shown in Table 7, there were differences in the time needed for different networks to
the efficiency of the deep-ocular model in terms of classification time. Meanwhile, as
extract features. In addition to accuracy, the AlexNet we used as feature extractor has the
shown in Table 7, there were differences in the time needed for different networks to extract
shortest running time (5.917 ± 0.001 s), which demonstrates the higher efficiency of our
features. In addition to accuracy, the AlexNet we used as feature extractor has the shortest
proposed model (AlexNet + ReliefF + XgBoost).
running time (5.917 ± 0.001 s), which demonstrates the higher efficiency of our proposed
Figure 7 illustrates the confusion matrix derived from the proposed model for the
model (AlexNet + ReliefF + XgBoost).
detection
Figure of7the three main
illustrates the classes of eye-related
confusion matrix deriveddiseases,
fromcataract, diabetic
the proposed retinopathy
model for the
and glaucoma, compared to normal retina. Additionally, as demonstrated in
detection of the three main classes of eye-related diseases, cataract, diabetic retinopathy Table 7, the
proposed
and model
glaucoma, in this study
compared performed
to normal retina.significantly
Additionally, better than the current
as demonstrated models
in Table in
7, the
terms of efficiency, classifying the test set in just 5.917 ± 0.001 s (n = 300),
proposed model in this study performed significantly better than the current models in thanks to its
smallerof neural
terms network
efficiency, construction.
classifying Ourinmodel
the test set can ±
just 5.917 drastically
0.001 s (nreduce
= 300),training
thanks totime
its
while assuring accuracy. The AUC values for our model (given in Figure
smaller neural network construction. Our model can drastically reduce training time while 8) improved
generally,accuracy.
assuring as can be observed.
The In comparison
AUC values for our modelto previous
(given inmethods,
Figure 8) the ensemble
improved model
generally,
suggested in this work can generally identify eye diseases with more
as can be observed. In comparison to previous methods, the ensemble model suggestedefficiency and accu-in
racy.
this work can generally identify eye diseases with more efficiency and accuracy.
Figure 8. AUC of a proposed model to detect glaucoma and cataract eye-related disease.
Figure 8. AUC of a proposed model to detect glaucoma and cataract eye-related dis
4.3. Feature Visualization through AI Interpretability
Grad-CAM,
4.3. Feature which standsthrough
Visualization for gradient
AI class activation map, is a technique used to
Interpretability
visualize the regions of an image that a deep neural network focuses on when making a
Grad-CAM,
particular prediction. which stands for
This visualization helpsgradient
to provideclass activation
insights map,
into what parts is a techn
of the
input imagethe
visualize contribute
regionsmostoftoantheimage
network’s decision-making
that a deep neural process. By highlighting
network focuses on w
these regions, Grad-CAM offers a better understanding of the features that the network
is recognizing. To integrate Grad-CAM into a model like our deep-ocular system, a few
steps are involved. First, you load your pre-trained model (in this case, using a modified
AlexNet) and the image you want to visualize. The model should have a global average
pooling layer followed by a dense layer for classification. Figure 9 shows the Grad-CAM
visualization of features classified using the proposed deep-ocular system.
The process began by obtaining the index of the predicted class. Then, we calculated
the gradients of the output score for the predicted class with respect to the final convo-
lutional layer’s output. These gradients indicate the importance of different parts of the
convolutional layer’s output in making the specific prediction. The higher the gradient, the
more influential that part of the image is in predicting the chosen class. After obtaining
the gradients, we calculated the heatmap by averaging the gradients along the channels
of the convolutional layer’s output. This heatmap was then normalized and resized to
match the dimensions of the original input image. The heatmap was further enhanced by
applying a color map, often the “jet” color map. Finally, the heatmap was overlaid on the
original image using an additive blending approach. The resulting image shows a visual
representation of where the model’s attention was focused when making the prediction.
input image contribute most to the network’s decision-making process. By highlighting
these regions, Grad-CAM offers a better understanding of the features that the network is
recognizing. To integrate Grad-CAM into a model like our deep-ocular system, a few steps
are involved. First, you load your pre-trained model (in this case, using a modified
AlexNet) and the image you want to visualize. The model should have a global average
Diagnostics 2023, 13, 3165 20 of 25
pooling layer followed by a dense layer for classification. Figure 9 shows the Grad-CAM
visualization of features classified using the proposed deep-ocular system.
9. A
Figure 9. A visual
visual Grad-CAM
Grad-CAMofofthetheproposed
proposeddeep-ocular
deep-ocularsystem
systemononretinographs, where
retinographs, where(a) figure
repre-
(a) represents
sents diabetes,diabetes,
(b) showsfigure (b) showsand
the glaucoma, the (c)
glaucoma,
indicatesand
the figure (c)eye-related
cataract indicates the cataract eye-re-
diseases.
lated diseases.
5. Discussion
Globally, thebegan
The process prevalence of eye problems
by obtaining the indexvaries
of the greatly according
predicted to age,
class. Then, wegender,
calculated oc-
cupation, lifestyle,
the gradients of theeconomic status,
output score for hygiene, habits,
the predicted andwith
class traditions.
respectAccording
to the finaltoconvolu-
a study
conducted onoutput.
tional layer’s people These
in tropical and temperate
gradients areas,
indicate the unavoidable
importance eye infections
of different parts ofare
themore
con-
common
volutionalinlayer’s
low-altitude
outputpopulations
in making the because of prediction.
specific factors suchThe
as dust,
higherhumidity, sunlight,
the gradient, the
and
moreother naturalthat
influential conditions [5].image
part of the Furthermore, ocular problems
is in predicting the chosenmanifest differently
class. After obtaining in
underdeveloped and developed cultures. Many underdeveloped nations,
the gradients, we calculated the heatmap by averaging the gradients along the channelsparticularly in
Asia, have true rates of ocular morbidity that are underdiagnosed and
of the convolutional layer’s output. This heatmap was then normalized and resized toneglected [6]. A
total of 285 million people are thought to have visual impairments worldwide,
match the dimensions of the original input image. The heatmap was further enhanced by of which
246 million have poor vision and 39 million are blind [7]. According to the World Health
Organization (WHO), 2.2 billion people worldwide suffer from nearsightedness or far-
sightedness [8]. According to estimates, 50% of these circumstances might have been
prevented or resolved [9]. Cataracts, age-related macular degeneration, glaucoma, diabetic
retinopathy, corneal opacity, trachoma, hypertension, and other disorders are among the
main causes of visual impairment [10].
Deep learning-based algorithms are increasingly being used in medical-picture anal-
ysis [15]. Several tasks, including identifying illnesses, have shown the effectiveness of
deep-learning-based models [16]. Automated illness diagnosis is one of the most crucial
Diagnostics 2023, 13, 3165 21 of 25
The integrated module successfully and precisely identified and classified eye-related
diseases. The proposed model (improved AlexNet, ReliefF, and XgBoost) had the highest
accuracy value, demonstrating its excellence. The feature extraction component had the
greatest impact on the model’s accuracy, with InceptionV3, SqueezeNet, and AlexNet
being the most accurate. Finding a suitable network for feature extraction is crucial for
successful classification. This study also compared the performance of five existing models
with the new model to confirm its usefulness. The proposed model had the best overall
accuracy rate (95.642%), reducing the training time and ensuring the accuracy, as reported
in Table 6 and Figure 7. The ensemble model improves our overall efficiency and accuracy
in identifying eye diseases compared to previous methods.
Utilizing the ODIR and RFMiD datasets, our approach integrated cutting-edge image
classification techniques through an ensemble-based transfer learning approach. Central
to our proposal was a novel three-step hybrid ensemble model, encompassing a classi-
fier, feature extractor, and feature selector. Our approach harnessed the potential of an
improved AlexNet architecture, featuring attention and dense layers, to enhance feature
extraction, adaptability, interpretability, and accuracy. This architecture demonstrated supe-
rior performance compared to other transfer learning approaches, especially for retinograph
classification tasks.
Moreover, our framework incorporates the ReliefF feature selection method to retain
the most salient features and reduce dimensionality. The subsequent classification using
XgBoost is capable of accurately identifying different ocular diseases. By integrating data
augmentation techniques, we effectively address class-imbalance concerns. Notably, our
deep-ocular model, predominantly built upon the AlexNet-ReliefF-XgBoost ensemble,
achieved an impressive accuracy of 95.13%. This underscores the potential of our proposed
approach to significantly aid dermatologists in making informed decisions for the early
diagnosis and screening of eye-related diseases, ultimately contributing to improved patient
care and outcomes.
(6) The proposed ensemble model, named “Deep-ocular,” achieved an impressive accu-
racy of 95.13%. This accuracy suggests that our system is highly capable of accurately
classifying various ocular diseases.
(7) The high accuracy achieved by our model indicates its potential to be a valuable tool
for dermatologists and healthcare professionals, aiding them in making early and
accurate decisions for diagnosing and screening eye-related diseases.
Table 8. Limitations of proposed system, when further developing and deploying the deep-ocular
system in clinical or research settings.
Limitation Description
The system’s performance might degrade when applied to datasets from different sources or
Limited Generalization populations due to differences in imaging quality, patient demographics, and diseases. In addition,
other ocular diseases such as hypertension should be considered.
The system’s performance heavily relies on the availability and diversity of high-quality labeled
Data Availability
retinal fundus images for training. Limited data could hinder its effectiveness.
Despite enhancements, overfitting might still occur due to the complexity of deep learning models.
Overfitting
Regularization techniques might be needed to mitigate this issue.
The system’s applicability is limited to eye diseases that can be diagnosed using retinal fundus
Limited to fundus images
images. Conditions that require other types of tests will not be detected.
Training and utilizing deep learning models can demand significant computational resources, which
Computational Resources
might limit the system’s accessibility in resource-constrained settings.
6. Conclusions
In conclusion, the challenges faced by clinicians and less-experienced ophthalmologists
in identifying early eye-related diseases have underscored the need for automated ocular
disease detection systems. Manual diagnosis is riddled with labor-intensive efforts and
the potential for errors, magnified by the diversity of ocular conditions such as glaucoma,
diabetic retinopathy, cataract, and normal cases. In response, this study presents a compre-
hensive solution in the form of an intelligent detection system for retinal fundus images.
Leveraging the power of deep learning algorithms, particularly the iAlexNet architecture
enhanced with attention and dense layers, offers a paradigm shift in feature extraction,
adaptability, interpretability, and accuracy within the realm of retinograph classification.
To realize this solution, this study harnessed the potential of ensemble-based transfer
learning, establishing a three-step hybrid ensemble model that seamlessly integrated a clas-
sifier, a feature extractor, and a feature selector. The initial phase involved feature extraction
utilizing a pre-trained AlexNet model, tailored with attention and dense layers to heighten
its performance. This improved architecture, iAlexNet, offers distinct advantages over
other transfer learning models, catering to the unique demands of retinograph classification
tasks. The feature selection phase employs ReliefF to identify and retain relevant features,
while the dimensionality is minimized for computational efficiency. Ultimately, a robust
XgBoost classifier culminates in disease classification, yielding differentiated outcomes for
various ocular conditions.
The efficacy of the proposed model was evidenced by the achievement of a remarkable
95.13% accuracy in the deep-ocular model. These promising results underscore the potential
of this ensemble approach, affirming its viability as a valuable tool for aiding healthcare
professionals in the timely diagnosis and screening of eye-related diseases. By combining
Diagnostics 2023, 13, 3165 24 of 25
cutting-edge techniques in deep learning, feature selection, and classification, this study
contributes to a pivotal advancement in ocular disease detection and management.
Author Contributions: Conceptualization, Q.A., M.A., A.A., R.S.A. and A.H.; data curation, Q.A.,
R.S.A. and A.H.; formal analysis, Q.A., M.A., A.A., R.S.A. and A.H.; funding acquisition, M.A. and
A.A.; investigation, Q.A., M.A., A.A. and R.S.A.; methodology, Q.A., M.A., R.S.A. and A.H.; project
administration, Q.A. and M.A.; resources, M.A., A.A. and A.H.; software, Q.A., M.A., A.A., R.S.A.
and A.H.; supervision, A.H.; validation, Q.A., M.A., A.A. and R.S.A.; visualization, R.S.A. and A.H.;
writing—original draft, Q.A., M.A., A.A., R.S.A. and A.H.; writing—review and editing, Q.A., M.A.,
A.A., R.S.A. and A.H. All authors have read and agreed to the published version of the manuscript.
Funding: This work was supported and funded by the Deanship of Scientific Research at Imam
Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-RP23079).
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The Python code is provided on GitHub (https://github.com/Qaisar2
56/Deep-ocular) to better understand the deep-ocular system.
Acknowledgments: This work was supported and funded by the Deanship of Scientific Research at
Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-RP23079).
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Abbas, Q.; Qureshi, I.; Yan, J.; Shaheed, K. Machine learning methods for diagnosis of eye-related diseases: A systematic review
study based on ophthalmic imaging modalities. Arch. Comput. Methods Eng. 2022, 29, 3861–3918. [CrossRef]
2. Sajid, M.Z.; Qureshi, I.; Abbas, Q.; Albathan, M.; Shaheed, K.; Youssef, A.; Hussain, A. Mobile-HR: An Ophthalmologic-Based
Classification System for Diagnosis of Hypertensive Retinopathy Using Optimized MobileNet Architecture. Diagnostics 2023,
13, 1439. [CrossRef] [PubMed]
3. Qureshi, I.; Abbas, Q.; Yan, J.; Hussain, A.; Shaheed, K.; Baig, A.R. Computer-Aided Detection of Hypertensive Retinopathy
Using Depth-Wise Separable CNN. Appl. Sci. 2022, 12, 12086. [CrossRef]
4. Guan, H.; Zhang, Y.; Wang, Z.; Chen, X.; Xue, J.; Ding, Y. The prevalence and county-level socioeconomic correlates of visual impairment
and glasses ownership among rural students in Western China. Front. Public Health 2023, 11, 1069793. [CrossRef] [PubMed]
5. Qureshi, I.; Ma, J.; Abbas, Q. Diabetic retinopathy detection and stage classification in eye fundus images using active deep
learning. Multimed. Tools Appl. 2021, 80, 11691–11721. [CrossRef]
6. Katibeh, M.; Pakravan, M.; Yaseri, M.; Pakbin, M.; Soleimanizad, R. Prevalence and causes of visual impairment and blindness in
central Iran; the Yazd eye study. J. Ophthalmic Vis. Res. 2015, 10, 279–285. [PubMed]
7. Ivleva, S.; Hajimajeed, M.; Kartashova, O.; Mikhaylov, A. A comparative study of regional variations in common visual
impairments among school-age children. Electron. J. Gen. Med. 2023, 20, em504. [CrossRef]
8. Ruamviboonsuk, P.; Tiwari, R.; Sayres, R.; Nganthavee, V.; Hemarat, K.; Kongprayoon, A.; Webster, D.R. Real-time diabetic
retinopathy screening by deep learning in a multisite national screening programme: A prospective interventional cohort study.
Lancet Digit. Health 2022, 4, e235–e244. [CrossRef] [PubMed]
9. Virgili, G.; Parravano, M.; Petri, D.; Maurutto, E.; Menchini, F.; Lanzetta, P.; Lucenteforte, E. The association between vision
impairment and depression: A systematic review of population-based studies. J. Clin. Med. 2022, 11, 2412. [CrossRef]
10. Müller, K.; Engel, C.; Loitsch, C.; Stiefelhagen, R.; Weber, G. Traveling more independently: A study on the diverse needs and
challenges of people with visual or mobility impairments in unfamiliar indoor environments. ACM Trans. Access. Comput.
TACCESS 2022, 15, 13. [CrossRef]
11. Ting, D.S.; Peng, L.; Varadarajan, A.V.; Keane, P.A.; Burlina, P.M.; Chiang, M.F.; Schmetterer, L.; Pasquale, L.R.; Bressler, N.M.;
Webster, D.R.; et al. Deep learning in ophthalmology: The technical and clinical considerations. Prog. Retin. Eye Res. 2019,
72, 100759. [CrossRef] [PubMed]
12. Fan, R.; Alipour, K.; Bowd, C.; Christopher, M.; Brye, N.; Proudfoot, J.A.; Zangwill, L.M. Detecting Glaucoma from Fundus
Photographs Using Deep Learning without Convolutions: Transformer for Improved Generalization. Ophthalmol. Sci. 2023,
3, 100233. [CrossRef]
13. He, J.; Li, C.; Ye, J.; Qiao, Y.; Gu, L. Self-speculation of clinical features based on knowledge distillation for accurate ocular disease
classification. Biomed. Signal Process. Control. 2021, 67, 102491. [CrossRef]
14. Zang, P.; Hormel, T.T.; Hwang, T.S.; Bailey, S.T.; Huang, D.; Jia, Y. Deep-Learning–Aided Diagnosis of Diabetic Retinopathy,
Age-Related Macular Degeneration, and Glaucoma Based on Structural and Angiographic OCT. Ophthalmol. Sci. 2023, 3, 100245.
[CrossRef] [PubMed]
Diagnostics 2023, 13, 3165 25 of 25
15. Roy, A.G.; Conjeti, S.; Karri, S.P.K.; Sheet, D.; Katouzian, A.; Wachinger, C.; Navab, N. ReLayNet: Retinal layer and fluid
segmentation of macular optical coherence tomography using fully convolutional networks. Biomed. Opt. Express 2017, 8,
3627–3642. [CrossRef]
16. Lee, C.S.; Tyring, A.J.; Deruyter, N.P.; Wu, Y.; Rokem, A.; Lee, A.Y. Deep-learning based, automated segmentation of macular
edema in optical coherence tomography. Biomed. Express 2017, 8, 3440–3448. [CrossRef]
17. Karri, S.P.K.; Chakraborty, D.; Chatterjee, J. Transfer learning based classification of optical coherence tomography images with
diabetic macular edema and dry age-related macular degeneration. Biomed. Opt. Express 2017, 8, 579–592. [CrossRef]
18. Oda, M.; Yamaguchi, T.; Fukuoka, H.; Ueno, Y.; Mori, K. Automated Eye Disease Classification Method from Anterior Eye Image
Using Anatomical Structure Focused Image Classification Technique. In Medical Imaging 2020: Computer-Aided Diagnosis; SPIE:
Houston, TX, USA, 2020; Volume 11314, pp. 991–996.
19. Eperjesi, F.; Fowler, C.W.; Kempster, A.J. Luminance and chromatic contrast effects on reading and object recognition in low
vision: A review of the literature. Ophthalmic Physiol. Opt. 1995, 15, 561–568. [CrossRef]
20. Ocular Disease Recognition Dataset. Available online: https://www.kaggle.com/andrewmvd/ocular-disease-recognition-odir5k
(accessed on 12 January 2022).
21. Pachade, S.; Porwal, P.; Thulkar, D.; Kokare, M.; Deshmukh, G.; Sahasrabuddhe, V.; Mériaudeau, F. Retinal fundus multi-disease
image dataset (rfmid): A dataset for multi-disease detection research. Data 2021, 6, 14. [CrossRef]
22. Tayal, A.; Gupta, J.; Solanki, A.; Bisht, K.; Nayyar, A.; Masud, M. DL-CNN-based approach with image processing techniques for
diagnosis of retinal diseases. Multimed. Syst. 2022, 28, 1417–1438. [CrossRef]
23. Meng, X.; Xi, X.; Yang, L.; Zhang, G.; Yin, Y.; Chen, X. Fast and effective optic disk localization based on convolutional neural
network. Neurocomputing 2018, 312, 285–295. [CrossRef]
24. Reddy, V.P.C.; Gurrala, K.K. Joint DR-DME classification using deep learning-CNN based modified grey-wolf optimizer with
variable weights. Biomed. Signal Process. Control. 2022, 73, 103439. [CrossRef]
25. Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.;
Cuadros, J.; et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus
photographs. JAMA 2016, 316, 24–35. [CrossRef]
26. Khan, M.S.M.; Ahmed, M.; Rasel, R.Z.; Khan, M.M. Cataract detection using convolutional neural network with VGG-19 model.
In Proceedings of the 2021 IEEE World AI IoT Congress (AIIoT), Virtual, 10–13 May 2021; pp. 0209–0212.
27. Li, N.; Li, T.; Hu, C.; Wang, K.; Kang, H. A benchmark of ocular disease intelligent recognition: One shot for multi-disease detection.
In Benchmarking, Measuring, and Optimizing: Proceedings of the Third BenchCouncil International Symposium, Bench 2020, Virtual Event,
15–16 November 2020; Revised Selected Papers 3; Springer International Publishing: Cham, Switzerland, 2021; pp. 177–193.
28. Sarki, R.; Ahmed, K.; Wang, H.; Zhang, Y.; Wang, K. Convolutional neural network for multi-class classification of diabetic eye
disease. EAI Endorsed Trans. Scalable Inf. Syst. 2021, 9, 23–38. [CrossRef]
29. Chea, N.; Nam, Y. Classification of Fundus Images Based on Deep Learning for Detecting Eye Diseases. Comput. Mater. Contin.
2021, 67, 412–426. [CrossRef]
30. Sarki, R.; Ahmed, K.; Wang, H.; Zhang, Y. Automated detection of mild and multi-class diabetic eye diseases using deep learning.
Health Inf. Sci. Syst. 2020, 8, 32. [CrossRef]
31. Triwijoyo, B.K.; Sabarguna, B.S.; Budiharto, W.; Abdurachman, E. Deep learning approach for classification of eye diseases based
on color fundus images. In Diabetes and Fundus OCT; Elsevier: Amsterdam, The Netherlands, 2020; pp. 25–57.
32. Badah, N.; Algefes, A.; AlArjani, A.; Mokni, R. Automatic eye disease detection using machine learning and deep learning
models. In Pervasive Computing and Social Networking: Proceedings of ICPCSN 2022, Salem, India, 3–4 March 2022; Springer Nature:
Singapore, 2022; pp. 773–787.
33. Elsawy, A.; Eleiwa, T.; Chase, C.; Ozcan, E.; Tolba, M.; Feuer, W.; Abdel-Mottaleb, M.; Shousha, M.A. Multidisease deep learning
neural network for the diagnosis of corneal diseases. Am. J. Ophthalmol. 2021, 226, 252–261. [CrossRef] [PubMed]
34. Sengar, N.; Joshi, R.C.; Dutta, M.K.; Burget, R. EyeDeep-Net: A multi-class diagnosis of retinal diseases using deep neural
network. Neural Comput. Appl. 2023, 35, 10551–10571. [CrossRef]
35. Chellaswamy, C.; Geetha, T.S.; Ramasubramanian, B.; Abirami, R.; Archana, B.; Bharathi, A.D. Optimized Convolutional Neural
Network based Multiple Eye Disease Detection and Information Sharing System. In Proceedings of the 2022 6th International
Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 25–27 May 2022; pp. 1105–1113. [CrossRef]
36. He, J.; Li, C.; Ye, J.; Qiao, Y.; Gu, L. Multi-label ocular disease classification with a dense correlation deep neural network. Biomed.
Signal Process. Control 2020, 63, 102167. [CrossRef]
37. Ocular Disease Intelligent Recognition through Deep Learning Architectures. Available online: https://jordicorbilla.github.io/
ocular-disease-intelligent-recognition-deep-learning/ (accessed on 12 January 2023).
38. Guergueb, T.; Akhloufi, M.A. Ocular diseases detection using recent deep learning techniques. In Proceedings of the Annual
International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Virtual, 1–5 November 2021; pp. 3336–3339.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.