Professional Documents
Culture Documents
anish_p140084ch
anish_p140084ch
anish_p140084ch
Abstract
Keywords: tumor, Magnetic Resonance Imaging, convolutional neural network, dense block
I. Introduction
Brain tumor is one of the dreadful diseases since several decades. There are several
aids for curing brain tumor. The abnormal cells in brain are termed as Gliomas. It is divided
into Low Grade Glioma (LGG) and High Grade Glioma (HGG). Tumor starts with less area
(LGG) and grows severe (HGG) which leads to death. Though the disease is severe, it can be
diagnosed when it is treated earlier. Medical tools are growing day by day which increases
the life of person. One such tool is the Magnetic Resonance Imaging (MRI) which scans the
brain using magnetic field and computer-generated radio waves. This helps the physician in
identifying tumor region and its severity.
The MRI images are available in three dimensions similar to our brain. The BRATS
dataset consists of four different modalities for a single patient. Classifying tumor in the
multimodal 3D data has several challenges. The location, size and shape of the tumour vary
from brain to brain. Automatic segmentation of tumor should identify the tumour area very
accurately. The computer aided automatic segmentation depends on deep learning nowadays.
As deep learning is a supervised learning, the system has to be trained with more number of
data.
As already mentioned, the brain image of a single patient consists of four modalites
and each modality has three dimensions. Training such a huge data using simple CNN itself
suffers from high computation time. The deep learning model used for segmentation suffers
from high computation time due to two important reasons.
The remaining of the chapter is organized as follows: Section 2 gives very recent
algorithms in brain tumor segmentation and classification. Section 3 elaborates the proposed
methods with all its phases. Section 4 demonstrates the proposed method with some
experiments. It also analyses the results with some comparison. Section 5 concludes the
work.
The multi modal 3D input images in BRATS 2018 consist of four modalities: T1,
T1c, T2 and FLAIR. The size of each input is 240 x 240 x 155. These four images are often
referred to as one modality of MRI, respectively, which can play different roles in the
segmentation of tumors. For example, the whole tumor segmentation is better performed with
Flair, and the segmentation of tumor core can be better segmented under T2. There are two
ways to fuse these modalities: Early Fusion and late fusion. In early fusion, the modalities on
low-level features are combined. In late fusion, each modality is merged with others in a deep
layer after an independent Convolutional Neural Network (CNN). Processing these images
takes very long time. There are several pre-processing methods to reduce the size of the
input. This paper identifies the Region of Interest from the MRI and then fed to the UNet
with LSTM model for segmentation. The architecture is shown in Fig. 1. It consists of Slice
Trimming (Inter-slice reduction) and Region Of Interest (ROI) Segmentation (Inter-slice
reduction). The average processing time of LSTM is 5x times faster than CNN model [17].
Hence, LSTM layer is to UNet model in this work.
The correlation between nearer slices paves the way to this section. The number of
slices in the BRATS dataset is 155. The slices are reduced by taking the average of nearby 5
slices. Let the number of slices in the 3D input image be
I t={S a 1 , S a 2 , . .. , S m } (2)
where
S1 + S2 +S 3 + S4 + S 5 S +S +S +S +S S +S + S +S +S
Sa 1 = , Sa 2= 6 7 8 9 10 , . . . , Sm = 151 152 153 154 155
5 5 5
155
The value of m= =31. The number of slices is reduced from 155 to 31. The value
5
of 5 is taken, because 155 is divisible by 5 and 31.
The next step is to find the ROI. In Brain tumor segmentation, the ROI denotes the
tumourous area in the brain. The two important challenges in identifying ROI are:
The first challenge is resolved by identifying the center pixel location of the tumour
area from the FlAIR image. The second challenge is resolved by having common detection
window for all images. The algorithm for identifying ROI is given in Algorithm 1.
Identifying Center Pixel Location: To find the center pixel location of the tumour
area, the histogram equalized image I hist of Flair is calculated. From I hist , a binary image D
eq eq
is created using
D= {0 1 i>T
Otherwise
(3)
In the above equation, i is the pixel value and T is a threshold. By looking the values
in I hist , it is observed that the pixel values of tumour area are above 250. Hence the
eq
threshold value is set to 250. The center pixel location is identified by calculating the
minimum and maximum location of white pixels in D. The average of these locations is the
center location.
D= {0 1 Otherwise
i> 250
1.3 End
1.4 Find the minimum ( x min, y min) coordinate whose pixel value is 1 in D.
1.5 Find the maximum ( x max , y max ) coordinate whose pixel value is 1 in D.
Hence the ROI size is made common for all images. But the location depends on the
input image.
3D Multimodal
T1c MRI InputT2Data Flair
T1
Trimmed Slices
Mask
ROI
LSTM
After finding the ROI in each modal of 3D MRI, it is given LSTM network model. It
segments and classifies the labels of tumors.
The most effective network for brain tumor segmentation is UNet. Our approach uses
convolutional LSTM instead of recurrent convolutional layer in UNet model. Convolutional
LSTM is convolutional version of LSTM [18], and it deals sequential data. The Feedback
Unet with Convolutional LSTM layer architecture is shown in Fig. 3. The features extracted
in the first round are maintained in the cell, and the features in the second round are extracted
based on the maintained features. Three major changes are made in U-Net. The first change is
to do feedback the output of U-Net to input layer. The second change is concatenation of the
outputs at the first round to the outputs at the second round. The third change is the usage of
convolutional
Segmentation
Results
The input for the U-Net are the ROIs extracted in intra-slice reduction. The
probability map of each class is obtained from the softmax function at the final layer.
Segmentation results are obtained from those probability maps. In this network model, image
and probability maps are the input of network, and probability maps of each class are
obtained at the final layer. After that, the input image and probability maps obtained at the
first round are fed into the network again, and the probability maps obtained at the second
round are used as the final segmentation result.
This section analyses the proposed method with some experiments. Initially, the
datasets used for the proposed method is briefly discussed, followed by performance
measures and the results. Finally, the proposed method is compared with recent methods and
various analyses are done with ablation study.
The proposed method is tested on two publicly available BRATS 2017 and 2018
datasets [19]. There are 431 cases (both HGG and LGG) in 2017 dataset and 476 cases in
2018 dataset. There are 146 training samples in both the datasets. The testing samples of
2017 and 2018 dataset are 146 and 191 respectively. Some examples of BRATS 2017 and
2018 are shown in Fig. 3. As already mentioned, the BRATS dataset consists of 4 MRI
modules: T1, T1c, T2 and Flair. Annotations include 3 tumor sub-regions: the enhancing
tumor, the peritumoral edema, and the necrotic and non-enhancing tumor core. The
annotations were combined into 3 nested sub-regions: Whole Tumor (WT), Tumor Core (TC)
and Enhancing Tumor (ET). Fig. 4 shows some sample MRI images from BRATS dataset.
Measure Formula
TP+TN
Accuracy Acr = × 100
TP+TN + FP+ FN
TP
Sensitivity Sensitivity=
TP+ FN
TN
Specificity Specificity=
(FP+TN )
2 ×TP
DSC DSC=
FP+ ( 2 ×TP ) + FN
*TP- True Positive, TN - True Negative, FP - False Positive, FN – False Negative
Results Analysis
The whole implementation is done in Matlab 2020a. The results of the proposed
method are given in Table 2 for both datasets. For training UNet, we adopted the Adaptive
Moment Estimation (Adam) optimizer, with a batch size 2 and an initial learning rate 10 -4.
Dataset/
BRATS 2017 BRATS 2018
Measure
Accuracy (%) 97.24 92.8
Sensitivity (%) 97.45 91.25
Specificity (%) 96.87 90.4
DSC (%) 97.84 88.88
From the above table, it is observed that the proposed method achieves high accuracy,
sensitivity, specificity and DSC (which are above 90%) on both datasets.
The efficiency of the proposed method is evaluated by comparing the results of the
proposed method with recent methods [2, 11, 20 – 25]. All these methods use deep learning
for segmentation and classification.
From the Table 3, it is observed that the results of the proposed method is little higher
than other methods on both the datasets.
The objective of this paper is to reduce the computation burden of deep learning. In
this section, the computation time is analysed with and without inter-intra slice reduction.
The proposed method is tested on Intel Core i7 2.8 GHz machine with a GPU NVIDIA
GeForce GTX 1050. Table 4 displays the computation time of proposed method.
The computation time of the proposed method in the machine with the above-
mentioned configuration is 12s for BRATS 2017 dataset and 10s for BRATS 2018 dataset.
This speed is 3 times faster than simple U-Net model. Thus the proposed method is efficient
in computation time with satisfiable performance metrics values.
V. Conclusion