Professional Documents
Culture Documents
1 s2.0 S1746809421001567 Main PDF
1 s2.0 S1746809421001567 Main PDF
1 s2.0 S1746809421001567 Main PDF
A R T I C L E I N F O A B S T R A C T
Keywords: Glaucoma is an ocular disorder that affects the optic nerve and ultimately leads to partial or complete vision loss.
Glaucoma Hence, there is a strong need for early screening of glaucoma. Earlier diagnosis schemes mostly rely on hand
Convolution crafted feature engineering. On the other hand, the non-handcrafted feature extraction methods are generally
Non-handcrafted feature
designed with the support of gradient-based algorithms that suffer from critical problems like overfitting and
RCGA
ECNet
demand for a larger set of samples for effective training. To mitigate these issues, in this paper, we propose a
novel non-handcrafted feature extraction method termed as evolutionary convolutional network (ECNet) for
automated detection of glaucoma from fundus images. The proposed method includes various layers such as
convolutional, compression, rectified linear unit (ReLU), and summation layer which facilitate the extraction of
discriminative features. An evolutionary algorithm called real-coded genetic algorithm (RCGA) is employed to
optimize the weights at different layers. The ECNet is trained using a criteria that maximizes the inter-class
distance and minimizes the intra-class variance of different classes. The final feature vectors are then sub
jected to a set of classifiers such as K-nearest neighbor (KNN), backpropagation neural network (BPNN), support
vector machine (SVM), extreme learning machine (ELM), and kernel ELM (K-ELM) to select optimum performing
model. The experimental results on a dataset of 1426 fundus images (589 normal and 837 glaucoma) demon
strate that the ECNet model with SVM yielded the highest accuracy of 97.20% compared to state-of-the-art
techniques. The proposed model can aid ophthalmologists to validate their screening.
1. Introduction In practice, three tests such as optic nerve head (ONH) evaluation,
intraocular pressure (IOP) measurement, and function-based visual field
Glaucoma is one of the most rapidly growing eye diseases across the test are carried out to detect glaucoma clinically [1]. Among them, ONH
globe that causes irreversible vision loss. It is caused due to the elevated evaluation has been found to be most suitable and widely used by the
intraocular pressure which damages the optic nerve (disc). Hence, the glaucoma experts [7,8]. This evaluation is usually done manually with
major clinical prodromes appear in the optic disc region [1–4]. Based on the help of fundus images and the procedure is tedious, time-consuming
the estimation, it is observed that around 64.3 million individuals aged and required skilled supervision. Therefore, the design of automated
40–80 years were diagnosed with glaucoma in 2013 across the globe and glaucoma diagnosis systems has recently attracted notable attention to
is expected to affect around 112 million people by 2040 [5,6]. Early obtain faster and reliable decision.
detection is thus very imperative that can reduce the progress rate and The vertical cup to disc ratio (CDR) is a widely accepted clinical
avert the vision loss. parameter that is used for ONH assessment. The CDR value of a
* Corresponding author.
E-mail addresses: drnayak@ieee.org (D.R. Nayak), dibyasundar@ieee.org (D. Das), bmajhi@nitrkl.ac.in (B. Majhi), sulatha.bhandary@manipal.edu
(S.V. Bhandary), aru@np.edu.sg (U.R. Acharya).
https://doi.org/10.1016/j.bspc.2021.102559
Received 31 May 2020; Received in revised form 11 February 2021; Accepted 6 March 2021
Available online 16 March 2021
1746-8094/© 2021 Elsevier Ltd. All rights reserved.
D.R. Nayak et al. Biomedical Signal Processing and Control 67 (2021) 102559
glaucoma patient is relatively larger than the normal one. The example strongly relies on the number of training samples used and they tend to
fundus images of both categories and the location of the optic disc and perform poorly with limited training samples. Further, these methods
cup region for one sample (green dotted circle) are depicted in Fig. 1. consider the classification error to tune the parameters. It is worth
Over the past decades, plenty of segmentation methods based on CDR mentioning that these models may not always generate the best feature
are proposed for glaucoma screening [9–11]. These CDR based methods descriptors since their training stops after reaching a high threshold
bypass other parameters and depend strongly on the segmentation ac value. Such performance on the training set does not guarantee better
curacy, but suffer from low sensitivity problems. Later, many performance on unknown clinical data. The development of
multi-stage pipeline based machine learning models are proposed with non-handcrafted feature extraction models which can evaluate the
the use of various handcrafted features such as retinal nerve fiber layer quality of features derived is thus needed. A review of notable glaucoma
thickness parameters [12], texture and higher-order spectra (HOS) detection methods along with their performance is reported in Table 1.
features [13,14], wavelet energy features [15], fast Fourier transform, This paper proposes a novel non-handcrafted feature extraction
spline interpolation and principal component analysis (PCA) based
features [16], Radon transform (RT) based features [17], Gabor trans Table 1
form and entropy features [18], empirical wavelet transform (EWT) and Summary of existing automated glaucoma detection methods.
correntropy [19], morphological features [20,21], variational mode
Reference Method Classifier No. of Accuracy
decomposition (VMD), fractal dimension (FD) and entropy features images (%)
[22], texton and local configuration pattern (LCP) features [23],
[13] HOS and texture features Random 60 91.00
concatenated wavelet and EWT features [24], RT and modified census
forest
transformation (MCT) based spatial envelope features [25], clinical and [15] DWT and energy features SMO 60 93.33
multiresolution features [26], quasi-bivariate VMD based texture fea [20] Morphological features SVM 67 94.11
tures [27], and local binary patterns (LBP) [28]. The classifiers used [18] Gabor transform and SVM 510 93.10
entropy features
commonly in these models include random forest, support vector ma
[19] EWT and correntropy LS-SVM 505 96.67
chine (SVM), least squares SVM (LS-SVM), K-nearest neighbor (KNN), features
feedforward neural network (FNN) and Naive Bayes. The choice of [22] VMD, FD and entropy LS-SVM 488 94.79
suitable feature descriptor as well as the classification model is a major features
concern with these types of approaches. Further, many of these models [23] Texton and LCP features KNN 702 95.70
[26] Clinical and MLP 86 97.67
have been validated using a less number of samples.
multiresolution features
Due to its advantages over handcrafted feature learning, the non- [25] RT and MCT spatial SVM 1000 97.00
handcrafted feature learning models have gained increasing popularity envelope features
in recent years [29–32]. Limited works have been reported to date using [24] Concatenated DWT and SVM 505 83.57
these methods for screening glaucoma due to the limited availability of EWT features
[27] Quasi-bivariate VMD and LS-SVM 505 86.13
labeled images. For instance, in [2], and [4,33–35] and [36], convolu texture features
tional neural network (CNN) based methods are applied to fundus im [33] Deep CNN – 650 –
ages to detect glaucoma, whereas, a sparse autoencoder based method is [4] AlexNet SVM 650 –
presented in [37]. However, these methods used gradient-based algo [2] CNN – 1426 95.60
[37] Sparse autoencoder 1426 95.30
rithms to tune the model parameters. Their generalization performance
–
2
D.R. Nayak et al. Biomedical Signal Processing and Control 67 (2021) 102559
model which is designed by combining various feature representation proposed model where the input image convolves with a group of ker
layers and an evolutionary algorithm. The major contributions of the nels (filters) that results in various significant feature maps. It acts
proposed work are as follows. similarly to that of conventional CNN.
ReLU layer: The ReLU layer is placed after the convolutional and
• In an automated system, the design of non-handcrafted features with compression layer which helps to include non-linearity in the model.
fewer parameters is preferred to ensure better generalization. A hi This activation outputs input value if it is positive otherwise it outputs
erarchical feature learning method with a combination of convolu zero values.
tional, ReLU, compression, and summation layers is proposed. The Compression layer: The compression layer is introduced after the
suggested method demands fewer parameters compared to tradi convolutional layer to reduce the size of each feature map by multi
tional non-handcrafted approaches. plying with a weighted feature transformation matrix (also called
• To tune the parameters at different learnable layers, gradient-based kernel). The major advantage of including such a layer is to filter the
algorithms have been commonly used which have limitations of required features based on the learned weights. The compression pro
demanding a large set of samples, trapping at local optima, and cess can be simply treated as a 2D matrix multiplication. For a given
thereby resulting in poor generalization performance. To mitigate input volume of size M1 × N1 × Q1 and the weight matrix of size N1 × N1∗ ,
such issues, the proposed model makes use of a meta-heuristic the output volume of compression block is of size M1 × N1∗ × Q1 , where
optimization algorithm, namely real-coded genetic algorithm N1 ≪N∗1 .
(RCGA) and adopts a strategy based on maximizing the inter-class Summation and sigmoid activation layer: In the summation layer,
distance and minimizing intra-class variability to extract high level the output activation of the compression block is summed across the
features from the images directly. depth and is then flattened to produce a 1D vector for each image. The
• The effectiveness of the proposed model is evaluated on a dataset flattened output is finally applied to sigmoid activation to obtain the
containing 1426 retinal fundus images. Further, the existing models final feature vector.
based on both machine learning and deep learning have been vali
dated using the same dataset and the results are compared. 2.1. Proposed ECNet framework
The remainder of this paper is structured as follows. The detail The proposed ECNet framework involves one convolutional layer,
description of the model is presented in Section 2. Section 3 presents the one compression layer, two ReLU layers and one summation layer fol
dataset used, experimental settings, results and observation. Eventually, lowed by a sigmoid activation layer. In the convolutional layer, a group
the concluding remarks of this work are drawn in Section 4. of random filters are used to convolve with the image. This layer inputs a
fundus image I of size m × n × p (i.e., 64 × 64 × 3 in our work) and
2. Proposed methodology convolves with f (i.e., 32 in this work) filters of size x1 × y1 × z1 (i.e.,
5 × 5 ×3). The convolution is accomplished with stride 1 and padding,
The design of optimal features using retinal fundus images is still a and thus generates a stacked volume of size m × n × f (i.e.,
difficult task due to large variability in the images. The performance of 64 × 64 × 32). The ReLU activation is then applied on the result of the
an automated system strongly relies on the kind of features used. Many convolutional layer. The compression operation is then performed using
handcrafted features and limited non-handcrafted features have been a kernel of size x2 × y2 (i.e., 64 × 10) across depth f to obtain a lower
applied to detect glaucoma [3]. In non-handcrafted feature extraction dimensional volume of size m × y2 × f (i.e., 64 × 10 × 32). Here, the
schemes, the gradient descent algorithm has been extensively harnessed conditions such as x2 = n and y2 ≪ n should be satisfied. The responses of
as the parameter tuning strategy to learn features through error opti this layer are then summed across the depth to derive a single feature
mization. However, learning optimal features with fewer samples is no map of size m × y2 (i.e., 64 × 10). Then, the output of the preceding layer
longer attainable using these algorithms. Therefore, we proposed a is applied to a non-linear function namely, sigmoid to limit the feature
feature extraction technique that uses a meta-heuristic approach to tune values within a range. The output of the last layer is flattened and we
the learnable parameters. The proposed model includes two phases: eventually obtain a feature vector of size (m * y2) × 1 (i.e., 640 × 1). An
feature learning and classification. In the first phase, it focuses on overview of different layers of the proposed feature extractor and their
learning important features directly from the training images, while in responses are graphically illustrated in Fig. 3.
the second phase, these features are fed to various classifiers separately
to classify the unknown samples. The overall classification model is
2.2. Mathematical formulation for feature learning
depicted in Fig. 2.
The feature learning model includes various layers namely, a con
Our ECNet model has two layers: convolutional and compression
volutional layer, ReLU layer, compression layer, summation layer, and
layer that includes learnable parameters. The weights associated in
sigmoid layer.
these layers are randomly initialized and an optimization algorithm is
Convolutional layer: This layer is one of the vital layers of our
thereafter used to tune the parameters based on criteria that maximizes
Fig. 2. Overall block diagram of the proposed model for automated glaucoma detection.
3
D.R. Nayak et al. Biomedical Signal Processing and Control 67 (2021) 102559
the distance between the mean of different classes and minimizes the 2.3.1. Population initialization
variance of individual classes. Based on these conditions, we formulated The RCGA uses a set of random solutions known as population to find
the feature extraction task as a multi-objective optimization problem the best solution. Here, an individual solution comprises a set of weights
and is given by, that is defined as
( )
∑ω ∑ ω ∑ D ℙl (θ1 , θ2 ) = [θ11 , θ21 , …, θa1 , θ12 , θ22 , …, θb2 ] (4)
Maximize M(θ1 , θ2 ) = |μi,d − μj,d | (1)
where a = x1 * y1 * z1 * f, b = x2 * y2, and l = 1, 2, …, popsize.
i=1 j=i+1 d=1
ω
∑ ∑
D
Minimize V(θ1 , θ2 ) = σ2i,d (2) 2.3.2. Fitness evaluation
i=1 d=1 The purpose of using the optimization technique is to obtain a
network that can extract optimal features from raw fundus images for
where θ1 and θ1 indicate weights of convolutional and compression fil glaucoma detection. The mean distance and variance for all classes are
ters respectively, ω denoted number of classes, D denotes feature evaluated using (1) and (2). To satisfy the multi-objective criteria, two
dimension, σ2i and μi denote variance and mean of ith class for a feature types of fitness have been taken into consideration to select the best
set Fθ1,θ1 which is defined as performing parents using the roulette wheel selection technique and
f
∑ comparing new with old solutions. The comparison is required to in
θ1 ,θ2
Fi,j = [(Ii,j ⊗ θq1 ) × θ2 ] (3) crease intra-class distance and minimize the inter-class variance. The
q=1 mathematical details are presented in the following sections.
where j = 1, 2, …, Ni, i = 1, 2, …, ω, Ii,j denotes jth sample of ith class, 2.3.3. Selection
and Fi,j
θ1 ,θ2
indicates its corresponding features. The selection of parents is one of the major steps of RCGA as domi
nant parents lead to global optimum. For this purpose, the roulette
2.3. RCGA based multi-objective optimization wheel selection method has been employed in this work which needs a
probability value per individual to complete the selection process.
Meta-heuristic optimization techniques have been drawing signifi However, two optimization criteria are used in this study as stated in (1)
cant attention in the past few decades due to their advantages of not and (2). Hence, to obtain a single probability value that satisfy our ob
requiring gradient information, bypassing the local optima, and relying jectives, we designed a combined probability function as follows.
on simple concepts. Hence, they have been employed extensively in
Mℙl (θ1 , θ2 ) Vℙl (θ1 , θ2 )
tuning parameters of various machine learning algorithms. These meta- Oℙl (θ1 , θ2 ) = α ∗ ∑ω ℙi − β ∗ ∑ω ℙi (5)
M (θ , θ ) i=1 V (θ1 , θ2 )
heuristic algorithms are mainly designed based on natural or biological i=1 1 2
4
D.R. Nayak et al. Biomedical Signal Processing and Control 67 (2021) 102559
where u is spread factor and η is distribution index. Here, η is set to 2 and available at [43]. The overall flowchart of the optimization model is
the value of u varies randomly in range [0,2] for each crossover given in Fig. 4.
operation.
2.4. Classification
2.3.5. Mutation
Mutation helps to explore the solution space. In this work, we used After learning the weights using the multi-objective RCGA method
the normally distributed mutation [42]. The mutation operation applied from the training set, the final feature vectors are generated. The clas
on the newly generated offspring is expressed as sification is then performed using different standard classifiers such as
SVM, KNN, BPNN, extreme learning machine (ELM) and kernel ELM (K-
Cij = Cij + N(0, ν) (8) ELM) [44].
where N(.) indicates normal distribution with variance ν and mean 0. 3. Experiments and results
2.3.6. Solution replacement strategy In this section, we present the materials used and results obtained
To compare the newly generated offspring with the existing parents, from all the experiments. The efficacy of proposed ECNet model was
the following replacement strategy has been formulated. tested on a larger dataset and the results for various combinations of
{ optimization techniques and classifiers were computed. A comparative
Ci if (MCi − VCi > MPj − VPj )
Pnew
j = (9) analysis was also performed with the existing automated methods. We
Pj otherwise
executed all experiments on a machine with 48 GB RAM and Xenon
where Pnew represents jth parent chosen from the population and Ci 2.4 GHz processor.
j
denotes the newly generated offspring. The code of our model is publicly
5
D.R. Nayak et al. Biomedical Signal Processing and Control 67 (2021) 102559
6
D.R. Nayak et al. Biomedical Signal Processing and Control 67 (2021) 102559
Fig. 5. Confusion matrices obtained using different optimization techniques with SVM classifier: (a) RCGA, (b) DE, and (c) PSO.
Fig. 6. Convergence plots of different optimization algorithms. Fig. 7. Performance plot obtained by the model in different runs.
4. Conclusion the learnable layers. A set of classifiers was employed to carry out the
classification task. Experimental results and comparisons confirmed the
This paper proposed a novel non-handcrafted feature extraction superiority of ECNet+RCGA and SVM method over other state-of-the-art
technique called ECNet for efficient glaucoma detection using fundus techniques. The model obtained a higher performance with an accuracy
images. The model composed of a sequence of convolutional, of 97.20% and can hence aid ophthalmologists to validate their
compression, ReLU and summation layers to generate meaningful fea screening. In future, the developed model can be used to detect the
tures. A real-coded genetic algorithm was used to tune the parameters of various stages of glaucoma automatically.
7
D.R. Nayak et al. Biomedical Signal Processing and Control 67 (2021) 102559
Table 4 [8] J.E. Morgan, N.J.L. Sheen, R.V. North, Y. Choong, E. Ansari, Digital imaging of the
Performance evaluation of different models using 10-fold CV strategy. optic nerve head: monoscopic and stereoscopic analysis, Br. J. Ophthalmol. 89 (7)
(2005) 879–884.
Method ACC SEN SPE F1-score [9] J. Cheng, J. Liu, Y. Xu, F. Yin, D.W.K. Wong, et al., Superpixel classification based
optic disc and optic cup segmentation for glaucoma screening, IEEE Trans. Med.
RCGA and SVM 0.9804 0.9749 0.9881 0.9831 Imaging 32 (6) (2013) 1019–1032.
DE and SVM 0.9537 0.9438 0.9677 0.9599 [10] G.D. Joshi, J. Sivaswamy, S.R. Krishnadas, Optic disk and cup segmentation from
PSOand SVM 0.9453 0.9415 0.9508 0.9529 monocular color retinal images for glaucoma assessment, IEEE Trans. Med.
Imaging 30 (6) (2011) 1192–1205.
Bold signifies the best results which has been obtained by our proposed model. [11] J. Cheng, F. Yin, D.W.K. Wong, D. Tao, J. Liu, Sparse dissimilarity-constrained
coding for glaucoma screening, IEEE Trans. Biomed. Eng. 62 (5) (2015)
1395–1403.
Table 5 [12] M.-L. Huang, H.-Y. Chen, W.-C. Huang, Y.-Y. Tsai, Linear discriminant analysis and
Performance comparison with state-of-the-art automated glaucoma detection artificial neural network for glaucoma diagnosis using scanning laser polarimetry-
variable cornea compensation measurements in Taiwan Chinese population,
methods using the same dataset. Graefe’s Arch. Clin. Exp. Ophthalmol. 248 (3) (2010) 435–441.
Method Classifier ACC (%) [13] U.R. Acharya, S. Dua, X. Du, C.K. Chua, et al., Automated diagnosis of glaucoma
using texture and higher order spectra features, IEEE Trans. Inf. Technol. Biomed.
HOS and texture features [13] Random forest 84.58 15 (3) (2011) 449–455.
Gabor transform and entropy features [18] SVM 88.08 [14] M.R.K. Mookiah, U.R. Acharya, C.M. Lim, A. Petznick, J.S. Suri, Data mining
EWT and correntropy features [19] LS-SVM 90.89 technique for automated diagnosis of glaucoma using higher order spectra and
Concatenated DWT and EWT features [24] SVM 86.92 wavelet energy features, Knowl.-Based Syst. 33 (2012) 73–82.
Deep CNN [33] – 93.22 [15] S. Dua, U.R. Acharya, P. Chowriappa, S.V. Sree, Wavelet-based energy features for
AlexNet (Holistic) [4] SVM 88.32 glaucomatous image classification, IEEE Trans. Inf. Technol. Biomed. 16 (1) (2012)
VGG16 (Holistic) [4] SVM 92.06 80–87.
CNN [2] – 95.60 [16] R. Bock, J. Meier, L.G. Nyúl, J. Hornegger, G. Michelson, Glaucoma risk index:
automated glaucoma detection from color fundus images, Med. Image Anal. 14 (3)
Sparse autoencoder [37] – 95.30
(2010) 471–481.
Proposed ECNet SVM 96.37
[17] K.P. Noronha, U.R. Acharya, K.P. Nayak, R.J. Martis, S.V. Bhandary, Automated
Bold signifies the best results which has been obtained by our proposed model. classification of glaucoma stages using higher order cumulant features, Biomed.
Signal Process. Control 10 (2014) 174–183.
[18] U.R. Acharya, E.Y.K. Ng, L.W.J. Eugene, K.P. Noronha, L.C. Min, K.P. Nayak, S.
Credit author statement V. Bhandary, Decision support system for the glaucoma using Gabor
transformation, Biomed. Signal Process. Control 15 (2015) 18–26.
[19] S. Maheshwari, R.B. Pachori, U.R. Acharya, Automated diagnosis of glaucoma
Deepak Ranjan Nayak: Conceptualization, Methodology, Software, using empirical wavelet transform and correntropy features extracted from fundus
Writing – Reviewing and Editing, Visualization images, IEEE J. Biomed. Health Inform. 21 (3) (2017) 803–813.
Dibyasundar Das: Formal Analysis, Methodology, Software Vali [20] A. Issac, M.P. Sarathi, M.K. Dutta, An adaptive threshold based image processing
technique for improved glaucoma detection and classification, Comput. Methods
dation, Writing – Original Draft Preparation Programs Biomed. 122 (2) (2015) 229–244.
Banshidhar Majhi: Formal Analysis, Investigation, Writing – [21] J. Nayak, R. Acharya, P.S. Bhat, N. Shetty, T.-C. Lim, Automated diagnosis of
Reviewing and Editing, Supervision glaucoma using digital fundus images, J. Med. Syst. 33 (5) (2009) 337.
[22] S. Maheshwari, R.B. Pachori, V. Kanhangad, S.V. Bhandary, U.R. Acharya, Iterative
Sulatha V. Bhandary: Data Curation, Investigation, Supervision, variational mode decomposition based automated detection of glaucoma using
Validation fundus images, Comput. Biol. Med. 88 (2017) 142–149.
U. Rajendra Acharya: Investigation, Writing – Reviewing and [23] U.R. Acharya, S. Bhat, J.E. Koh, S.V. Bhandary, H. Adeli, A novel algorithm to
detect glaucoma risk using texton and local configuration pattern features
Editing, Supervision, Validation extracted from fundus images, Comput. Biol. Med. 88 (2017) 72–83.
[24] B.S. Kirar, D.K. Agrawal, Computer aided diagnosis of glaucoma using discrete and
Acknowledgments empirical wavelet transform from fundus images, IET Image Process. 13 (1) (2018)
73–82.
[25] U. Raghavendra, S.V. Bhandary, A. Gudigar, U.R. Acharya, Novel expert system for
This work is supported by the Science and Engineering Research glaucoma identification using non-parametric spatial envelope energy spectrum
Board (SERB), Department of Science and Technology, Govt. of India with fundus images, Biocybern. Biomed. Eng. 38 (1) (2018) 170–180.
[26] T. Kausu, V.P. Gopi, K.A. Wahid, W. Doma, S.I. Niwas, Combination of clinical and
under project No. SRG/2020/001460. Authors thank department of multiresolution features for glaucoma detection and its classification using fundus
Ophthalmology KMC Manipal for sharing the images for this work. images, Biocybern. Biomed. Eng. 38 (2) (2018) 329–341.
[27] D.K. Agrawal, B.S. Kirar, R.B. Pachori, Automated glaucoma detection using quasi-
bivariate variational mode decomposition from fundus images, IET Image Process.
Declaration of Competing Interest 13 (13) (2019) 2401–2408.
[28] S. Maheshwari, V. Kanhangad, R.B. Pachori, S.V. Bhandary, U.R. Acharya,
Automated glaucoma diagnosis using bit-plane slicing and local binary pattern
The authors report no declarations of interest. techniques, Comput. Biol. Med. 105 (2019) 72–80.
[29] G. Litjens, T. Kooi, B.E. Bejnordi, A.A.A. Setio, F. Ciompi, M. Ghafoorian, J.A. van
References der Laak, B. Van Ginneken, C.I. Sánchez, A survey on deep learning in medical
image analysis, Med. Image Anal. 42 (2017) 60–88.
[30] D.R. Nayak, R. Dash, B. Majhi, R.B. Pachori, Y. Zhang, A deep stacked random
[1] H. Fu, J. Cheng, Y. Xu, C. Zhang, D.W.K. Wong, J. Liu, X. Cao, Disc-aware ensemble
vector functional link network autoencoder for diagnosis of brain abnormalities
network for glaucoma screening from fundus image, IEEE Trans. Med. Imaging 37
and breast cancer, Biomed. Signal Process. Control 58 (2020) 101860.
(11) (2018) 2493–2501.
[31] Y. Guo, Y. Liu, A. Oerlemans, S. Lao, S. Wu, M.S. Lew, Deep learning for visual
[2] U. Raghavendra, H. Fujita, S.V. Bhandary, A. Gudigar, J.H. Tan, U.R. Acharya,
understanding: a review, Neurocomputing 187 (2016) 27–48.
Deep convolution neural network for accurate diagnosis of glaucoma using digital
[32] D.R. Nayak, R. Dash, B. Majhi, Automated diagnosis of multi-class brain
fundus images, Inf. Sci. 441 (2018) 41–49.
abnormalities using MRI images: a deep convolutional neural network based
[3] Y. Hagiwara, J.E.W. Koh, J.H. Tan, S.V. Bhandary, A. Laude, E.J. Ciaccio, L. Tong,
method, Pattern Recognit. Lett. 138 (2020) 385–391.
U.R. Acharya, Computer-aided diagnosis of glaucoma using fundus images: a
[33] X. Chen, Y. Xu, D.W.K. Wong, T.Y. Wong, J. Liu, Glaucoma detection based on deep
review, Comput. Methods Programs Biomed. 165 (2018) 1–12.
convolutional neural network, in: 37th Annual International Conference of the
[4] A. Li, J. Cheng, D.W.K. Wong, J. Liu, Integrating holistic and local deep features for
IEEE Engineering in Medicine and Biology Society (EMBC), IEEE, 2015,
glaucoma classification, in: 38th Annual International Conference of the IEEE
pp. 715–718.
Engineering in Medicine and Biology Society (EMBC), IEEE, 2016, pp. 1328–1331.
[34] X. Chen, Y. Xu, S. Yan, D.W.K. Wong, T.Y. Wong, J. Liu, Automatic feature learning
[5] Y.-C. Tham, X. Li, T.Y. Wong, H.A. Quigley, T. Aung, C.-Y. Cheng, Global
for glaucoma detection based on deep learning, in: International Conference on
prevalence of glaucoma and projections of glaucoma burden through 2040: a
MICCAI, Springer, 2015, pp. 669–677.
systematic review and meta-analysis, Ophthalmology 121 (11) (2014) 2081–2090.
[35] J.I. Orlando, E. Prokofyeva, M. del Fresno, M.B. Blaschko, Convolutional neural
[6] H.A. Quigley, A.T. Broman, The number of people with glaucoma worldwide in
network transfer for automated glaucoma identification, 12th International
2010 and 2020, Br. J. Ophthalmol. 90 (3) (2006) 262–267.
Symposium on Medical Information Processing and Analysis, vol. 10160 (2017)
[7] H. Fu, Y. Xu, S. Lin, X. Zhang, D.W.K. Wong, et al., Segmentation and
101600U.
quantification for angle-closure glaucoma assessment in anterior segment OCT,
IEEE Trans. Med. Imaging 36 (9) (2017) 1930–1938.
8
D.R. Nayak et al. Biomedical Signal Processing and Control 67 (2021) 102559
[36] Y. Chai, L. He, Q. Mei, H. Liu, L. Xu, Deep learning through two-branch [40] S. Yu, S.-Q. Kuang, Convergence and convergence rate analysis of elitist genetic
convolutional neuron network for glaucoma diagnosis, in: International algorithm based on martingale approach, Control Theory Appl. 27 (7) (2010)
Conference on Smart Health, Springer, 2017, pp. 191–201. 843–848.
[37] U. Raghavendra, A. Gudigar, S.V. Bhandary, T.N. Rao, E.J. Ciaccio, U.R. Acharya, [41] K. Deb, M. Goyal, A combined genetic adaptive search (GeneAS) for engineering
A two layer sparse autoencoder for glaucoma identification with fundus images, design, Comput. Sci. Inform. 26 (1996) 30–45.
J. Med. Syst. 43 (9) (2019) 299. [42] G. Rudolph, Asymptotical convergence rates of simple evolutionary algorithms
[38] J.R. Sampson, Adaptation in Natural and Artificial Systems (John H. Holland), under factorizing mutation distributions, in: European Conference on Artificial
Society for Industrial and Applied Mathematics, 1976. Evolution, Springer, 1997, pp. 223–233.
[39] D.E. Goldberg, Genetic Algorithms in Search, Optimization and Machine Learning, [43] CNet model code, https://github.com/Dibyasundar/ECNET_glaucoma.
Addison Wesley Publishing Co. Inc., 1989. [44] G.B. Huang, C.K. Siew, Extreme learning machine: RBF network case, in: 8th
International Conference on Control, Automation, Robotics and Vision, vol. 2,
IEEE, 2004, pp. 1029–1036.