Professional Documents
Culture Documents
Application of Residual Network Architecture On Covid-19 Chest X-Ray Classification
Application of Residual Network Architecture On Covid-19 Chest X-Ray Classification
Application of Residual Network Architecture On Covid-19 Chest X-Ray Classification
Abstract — Convolutional Neural Network (CNN) has training process. In this study, Imagenet weights are used on
proven with good performance in the area of feature extraction. Resnet with a number of convolution layers of 50 and 101.
Classification of medical images is often faced with the lack of The bottleneck structure in Resnet-50 and Resnet-101 serves
sufficient amount of data. Therefore, Transfer Learning can be to reduce the computational load due to an increase in
applied to overcome these problems. Chest x-ray data are weight on a deep network [3]. Model training procedures
complex and require deeper layers for specific features. Resnet that can be used include data splitting techniques,
built with deep layers specifically focuses on problems that Optimizers, and minor changes on hyperparameter setting
often occur in high-depth architectures, which are prone to
[4].
decreased accuracy and training errors. Some of the aspects are
able to affect the performance of the model such as the depth of The optimization algorithm used in the training process
convolution layers and training procedures, which include data can affect the speed and performance of deep learning
splitting technique and Optimizers. In this study, the Hold Out predictions [5] Adaptive Moment Estimation (Adam) and
data splitting and k-fold cross validation of 5 folds with Schocastic Gradient Descent (SGD) optimization
Optimizer Adaptive Moment Estimation (Adam) and algorithms are applied to minimize errors during the
Schocastic Gradient Descent (SGD) on the Resnet-50 and training process. The need for generalization speed can be
Resnet-101 architectures. The training procedure was applied
done by Adam Optimizer while SGD excels in
to 15143 Chest x-ray images measuring 224x224 pixels with
parameters epoch 50 and batch size 100. The best value was
convergence speed [6]. The Hold Out data splitting and k-
obtained using k-fold cross validation on Resnet-50 using the fold cross validation is part of the training procedure that
SGD optimizer with 99% accuracy. affects the generalizability of the model used [7].
Research conducted by Narin, et al has implemented the
Keywords — Resnet, Hold out, k-fold cross validation, Adam, Resnet-50 and 101 variants along with other non-residual
SGD. architectures on 7406 Chest x-ray data collected from 3
different sources, using the Adam Optimizer along with 2,
I. INTRODUCTION 5, and 10 folds cross validation and batch size 3, resulted in
Residual Network (Resnet) is a Neural Network Resnet-50 with an accuracy of 96.1%, 99.5% and 99.7%
architecture proposed by He, et al as research on the for each dataset [8]. Research conducted by Haq, et al on 2
correlation between the depth of the network layer with the datasets that have been augmented amounted to 12864,
effectiveness level of a Neural Network. The research came with the distribution of data Hold Out 70:30 resulting in
to the conclusion that a deep and centralized network can lead 95.20% accuracy and 94.87% on Resnet-50 obtained from
to a decrease in accuracy. This problem arises since more and 100 epochs. The Optimizer used is SGD with of 0.0001 as
more layers are added to the network architecture, the higher a learning rate parameter and a batch size of 100 [9].
the possibility of errors during training [1]. Resnet works Hamlili, et al [10] applied the 4-fold cross validation
effectively in overcoming the vanishing gradient and data splitting technique on binary and multi-class
exploding gradient issues [2] . classification using Transfer Learning Resnet-50 on 6173
Resnet performance may vary depending on the type of Chest x-ray datasets. Data augmentation was applied to
dataset and the type of improvisation carried out in the overcome the prediction imbalance. Training process is
carried out in 2 binary classification scenarios and 1 class 3 Fig. 2. Chest x-ray dataset [11]
classification scenario. The accuracy results obtained were
99.9% and 99.8% for binary classification, meanwhile the 3 B. Preprocessing
class classification resulted in 97.3% accuracy. The preprocessing stage consisted of the following
In this study, Transfer Learning was applied to the data procedures of resizing all the images into the same size of
of 15143 Chest x-ray images consisting of 3 classes, 224x224 pixels. In addition, the image size is increased with
namely Covid-19, Normal, and Pneumonia. The a 3x3 zero padding for each edge of the image. This serves to
augmentation process is done in the form of random maintain the original shape of the image when random
cropping. The training procedure was applied to Resnet-50 cropping augmentation is performed. In Deep Learning, the
and Resnet-101 in several scenarios, including using the number of datasets can be multiplied by the augmentation
Adam Optimizer and SGD Optimizer, with data splitting process. For medical datasets such as x-ray images, not all
Hold out and 5-fold cross validation. augmentations can be applied. Geometric augmentation such
II. METHOD as rotation and scaling has the potential to reduce accuracy
[12].
C. Resnet Architecture
zj
e
f j (z )=
∑ e zk (1)
k
The calculated neuron index is denoted as j, the A. Training Process
overall index neuron is denoted by k, and z is the neuron
The Resnet-50 and Resnet-101 training procedures use
from the output process. The number of neurons used in the
final classification process is 3 neurons. The workings of two data sharing techniques, namely Hold out and 5-fold
weight transfer and classification on the Chest x-ray dataset cross validation, as well as several hyperparameter settings
can be seen in Figure 4. including 50 epochs, batch size 100, and learning rate
0.00001. Data sharing Hold out divides data with the
formation of 70% : 30% [15], 80% : 20% [16], and 90% :
10% [17] while k-fold cross validation with a total of k=5
[8] equally divides the training dataset into 5 parts
randomly. The training process is carried out using a
Stratified k-fold so that each fold represents a complete
sample of each target class [18]. Based on the training
process carried out on the Adam Optimizer and SGD, the
accuracy results obtained from data sharing with Hold out
and 5-fold cross validation are represented by tables 1, 2, 3
and 4 as well as graphs 6 and 7.