Lung Cancer Detection and Prediction of Cancer Stages Using Image Processing

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 9

Lung Cancer Detection and Prediction of Cancer

Stages Using Image Processing


2021 3rd International Conference on Electrical, Control and Instrumentation Engineering (ICECIE) | 978-1-6654-4966-3/21/$31.00 ©2021 IEEE | DOI: 10.1109/ICECIE52348.2021.9664658

S.A.D.L.V. Senarathna S.P.Y.A.A. Piyumal R. Hirshan W.G.C.W. Kumara


Dept. of Electrical Dept. of Electrical Dept. of Electrical Dept. of Computer
and Telecommunication and Telecommunication and Telecommunication Science and
Engineering, Engineering, Engineering, Engineering,
South Eastern University South Eastern University South Eastern University South Eastern University
of Sri Lanka, of Sri Lanka, of Sri Lanka, of Sri Lanka,
Oluvil, Sri Lanka Oluvil, Sri Lanka Oluvil, Sri Lanka Oluvil, Sri Lanka
lakshithavimuth8@gmail.com spya.anushka@gmail.com rajehirshan@seu.ac.lk chinthakawk@seu.ac.lk

Abstract—Lung cancer is one of the most common and lung cancer (8.6 per 100,000 population) and it had been
dangerous cancers in the world. However, lives can be saved the sixth commonest cancer among all Sri Lankans (5.4 per
through early diagnosis by CT scan images, which is the best 100,000 population) [2]. The risk of lung cancer in Asian
imaging technique in the medical field for early treatment.
Though CT scan imaging is the best technique, doctors and countries is five times higher among smokers compared to
radiologists face some difficulties such as not being able to non-smokers. By early detection, the 5-year survival rate of
diagnose early and commence treatment and to interpret and patients with lung cancer may be improved by about 50% [3].
identify cancer from CT scan images because of the limita- Since the early clinical manifestations of lung cancer are not
tion of equipment and specialists. Therefore, to identify the obvious, early detection of lung cancer will encounter many
cancerous cells accurately, computer-aided diagnosis can be
more helpful for doctors. Computer-aided techniques based on difficulties, which results in the best treatment period for lung
image processing and machine learning have been extensively cancer being missed. Thus, early lung cancer detection and
researched, and are being implemented currently to address early treatment have become important issues. Lung cancer
this issue. This research is mainly focused on evaluating and is usually manifested as pulmonary nodules, and Computed
analyzing the different computer-aided techniques, find out their Tomography (CT) is an important means of examining lung
limitations and drawbacks and finally, propose a new model with
improvements. In the methodology section of this research, lung nodules. For lung CT images, hundreds of DICOM (Digital
cancer detection techniques were sorted and listed based on their Image Communication in Medicine) format images can be
detection accuracy. The techniques were analyzed on each step, produced at once [4]. Reading the hundreds of images takes
and overall limitations and disadvantages were pointed out. It many resources, while lesions often occupy only a tiny part of
is found that some techniques have low accuracy and some have an image. There is an enormous workload and misdiagnosis
higher accuracy, but not nearer to 100%. Therefore, our project
target is to make a lung cancer detection model using CT scan risk for artificial discrimination. Therefore, it is necessary
images with high accuracy and predicting the lung cancer stage. to introduce a more objective and intelligent computer-aided
detection technology (CAD) [5].
Index Terms—Cancer detection, Convolutional Neural Net-
work (CNN), Connected Component Labeling, Digital image Currently, radiologists and doctors usually perform visual
processing, Feature extraction, Hough Transform, Lung cancer, scanning of medical images to detect an abnormality in the
Lung cancer stage prediction.
human body. Hence, the procedure of diagnosing diseases
manually using medical images is a complex process. The
I. I NTRODUCTION categorization of disease will be difficult, and errors can
Lung cancer, the foremost common cancer in both men and occur by the blurring of anatomical structures surrounding
women, is a major disease burden. According to the World the lung area, the small size of lesions, and also the different
Health Organization (WHO), amount of new lung cancer experiences of the radiologist generate a different interpre-
cases is about 2.21 million and 1.8 million deaths, accounting tation. And also, current models which are based on image
for about 13% of all cancer diagnoses in 2020 worldwide. processing have extremely high false-positive rates in which
The mortality of lung cancer accounts for about 27% of all the result is incorrectly indicated as the presence of a disease
cancer deaths [1]. when the disease is not present, which do not allow oncolo-
According to the reports from the National Cancer Registry gists and radiologists to focus on patients with an imminent
of Sri Lanka, the second commonest cancer among males was cancer threat. Thus, by building a classification model that
reduces the number of false positives and false negatives,
patients can have earlier access to life-saving interventions
978-1-6654-4966-3/21/$31.00 ©2021 IEEE and allow doctors to prioritize patient care.

Authorized licensed use limited to: Zhejiang University. Downloaded on September 23,2023 at 02:21:04 UTC from IEEE Xplore. Restrictions apply.
Cancer prediction is the ability to predict the existence of A feature-wise comparison of the related work is provided in
cancer by recognizing a particular lesion or lesions in the CT Table I.
image with the background evidence of identified risk factors The general trend in radiology classification involves lung
and the favorable clinical picture of an individual. The stage cancer detection and classification for deep learning. Novel
of cancer can describe how big it is and whether it has spread. Fuzzy Particle Swarm Optimization Convolution Neural Net-
There are four stages. If nodule size d < 3 cm and it has not work (FPSOCNN) deep learning algorithm, which reduces
spread to the nearby lymph node means it is a level (I) lung the computational complexity of CNN was used to detect
cancer. If the range of nodule size d is 3 cm < d < 5 cm the cancerous lung nodules from the given input lung image
and it has not spread to the nearby lymph node means it and to classify the lung cancer and its severity. Fuzzy Particle
is a level (II) lung cancer. If the range of nodule size d is Swarm Optimization (FPSO) algorithm is applied for extract-
5 cm < d < 7 cm or any number and it has spread to the ing the features like texture, geometric, volumetric, and in-
nearby lymph node means it is a level (III) lung cancer. If tensity [17]. Moreover, to detect lung cancer, a hybrid model
nodule radius is any number and it has spread as more than based on the six different machine learning classifiers, three
one tumor outside the chest, such as to distant lymph nodes CNN models, and the mRMR feature selection method was
and/or to other organs such as the liver, bones, or brain, means proposed. Measuring the success of CNNs and the machine
it is a level (lV) lung cancer [6]. learning classifiers without image augmentation techniques
The organization of this paper is as follows. A literature was experimented using a combination of AlexNet, k-NN and
survey was done, which is related to our study, as explained in mRMR method provided the most effective results with an
Section II. We have discussed the proposed method in Section accuracy of 99.51%, the sensitivity of 99.32%, and specificity
III, which includes dataset preprocessing, cancer detection of 99.71% [18].
processes, and Hough transform-based cancer nodule stage The publicly available datasets are consisting of DICOM
prediction process. Experiment results and discussion of the image format, which is a standard protocol for the man-
model are explained in Section IV. The conclusion is given agement (produce, store, display, process) and transmission
in Section V. (share, send, query) of medical images and related data and
are used in many healthcare facilities. As per The Cancer
II. R ELATED W ORK
Imaging Archive (TCIA) (www.cancerimagingarchive.net),
In this chapter, a summary of the literature on the analysis which de-identifies and hosts an archive of medical images
of lung cancer is presented considering the dataset and of cancer accessible for public download, there are 32 lung-
methods. The efficiency of the CAD model depends on how related datasets out of 144 all cancer datasets [19]. We
the model performs with real-world problems. Spatial domain have studied the following randomly selected lung cancer
and frequency domain features technique was used to detect datasets, which filtered as cancer type: lung cancer, and
pulmonary tumor volume from positron emission tomography status: completed, are discussed in Table II.
images and the performance was analyzed by k-Nearest
Neighbor and Support Vector Machines (SVM) to propose III. P ROPOSED M ETHOD
a method to automatically delineate the tumor and detect The proposed method can be divided into three stages
its volume, which helps in accurate radiotherapy treatment as in the Fig. 1. Initially, the CT images which are in
planning. Moments, discrete cosine transforms and wavelets DICOM image format were converted into JPG file format.
of features were compared with different classifiers [7]. Apart In the image preprocessing stage, median filter and binary
from that, a Genetic algorithm and conventional template thresholding algorithm were used to segment the lungs from
matching methods were applied along the lung wall to detect the CT image. In the cancer detection stage, CNN architecture
the nodules in chest helical CT images by using nodular [20] was tested with the preprocessed CT scan image dataset.
models with Gaussian distribution. The proposed algorithm Finally, Fourier transforms [21], Gabor filter [22], and Otsu's
is a new template-matching algorithm for the detection of thresholding algorithms [23] were applied to cancer detected
nodules in chest helical CT images using modular models CT images to extract the nodules from the lung, and Hough
with Gaussian distribution as reference images [8]. Decision transform [24] was used to calculate the maximum possible
Tree, Support Vector Machines, Rule-based classification, outer diameter of the nodule, which is used to detect the stage
and Naive Bayes were used to extract the features from of the lung cancer, while the other researchers implemented
the heterogeneous dataset and predict non-small cell lung their models to classify the detected cancer as positive or
cancer [9]. Moreover, Bayesian classification and Hopfield negative. The details of each step are discussed in the
Neural Network algorithm were used to extract and segment following sections.
the sputum cells, which allow determining the classification The process of dividing a digital image into multiple
parameters [10]. A novel approach for predicting Lung cancer segments is called image segmentation. As an image seg-
nodule at an early stage using SVM Classifier has been menting algorithm, a binary thresholding algorithm was used
obtained with an accuracy of 95.12%, and a comparison of to convert input Igray into a binary image Ib . In binary
classification accuracy for ANN, KNN, and SVM Classifiers thresholding, two levels need to be assigned, and according
was made on lung CT scan images of stage I and stage II [11]. to the pixel values that are below or above the specified

Authorized licensed use limited to: Zhejiang University. Downloaded on September 23,2023 at 02:21:04 UTC from IEEE Xplore. Restrictions apply.
TABLE I. S UMMARY OF S IMILAR T ECHNOLOGIES
Title Final result Dataset used Algorithm Accuracy of the models (%)
KNN 92.00
[7] Detecting lung cancer Discovery STE8 CT scanner
SVM 90.00
[8] Detecting lung cancer Their dataset GA 72.00
DT 77.50
[9] Detecting lung cancer CT images from the Moffitt Cancer center
SVM 65.00
BC 64.91
[10] Detecting lung cancer Dataset of 88 scan images
HNN 88.62
KNN 85.87
[11] Detecting and predicting cancer stage I or II Lung Image Dataset Consortium (LIDC) SVM 95.12
ANN 92.68
Note: KNN - k-Nearest Neighbor, SVM - Support Vector Machines, ANN - Artificial Neural Network, GA - Genetic algorithm, DT - Decision Tree,
BC - Bayesian classification, HNN - Hopfield Neural Network.

TABLE II. L UNG C ANCER DATASET D ETAILS


Size
Format No. of
Of the No. of
Title of the partici- Modalities
Dataset images
images pants
(GB)
SPIE-AAPM Lung CT Challenge [12] 12.1 22,489 DICOM 70 CT
A Large-scale CT and PET/CT Dataset for Lung
127.2 251,135 DICOM 355 CT, PET
Cancer Diagnosis [13]
NSCLC-Radiomics-Interobserver [14][16] 3.2 3,886 DICOM 22 CT, RTSTRUCT, SEG
Lung Fused-CT Pathology [15][17] 5.5 11,210 DICOM 6 CT, Histology compartments mapped on CT
Lung CT Segmentation Challenge 2017 [16][18] 4.8 9,593 DICOM 60 CT, RT, RTSTRUCT

As the final step of image preprocessing, segmented image


Iseg was applied to the cancer detection model as described
below.
A. Cancer Detection
The second stage of the proposed system is to correctly
discriminate between normal and lung cancer images, and
CNN architecture which is based on supervised learning was
used to classify the repressed images. The proposed CNN
architecture for cancer detection is described in Table III,
which includes used layers, number of kernels, kernel size,
loss function used, and optimizer.

TABLE III. CNN A RCHITECTURE IN THE P ROPOSED M ETHOD


Loss
Layer No. of kernels Kernel size Optimizer
function
Conv2D 32 3×3
Conv2D 32 3×3
Maxpool2D 2×2 Sparse
Conv2D 64 3×3 Categorical
Conv2D 64 3×3 msprop
Cross-
Maxpool2D 2×2 entropy
Flatten
Dense 512
Fig. 1. Overview of the proposed lung cancer detection system.
Flatten 4

Layer view of the above-mentioned CNN architecture is


threshold levels, the corresponding output pixel is marked as shown in Fig. 2.
white (foreground) or black (background) respectively [25].
Canny edge detection algorithm, which is an inbuilt function B. Stage Prediction
was applied to Ib image and created a binarized mask, which The procedure of the cancer stage prediction on the cancer
was applied to the input image I to segment the lungs from detected image is explained below. It consists of seven steps
the CT image [26]. and four different algorithms.

Authorized licensed use limited to: Zhejiang University. Downloaded on September 23,2023 at 02:21:04 UTC from IEEE Xplore. Restrictions apply.
2
x0 + γ 2 y 0 2πx0
g(x, y, λ, θ, ψ, γ, δ) = exp(− 2
)sin(i( + ψ)),
2σ λ
(5)
where, x0 = xcos(θ) + ysin(θ), y 0 = ycos(θ) − xsin(θ), λ:
Wavelength of the sinusoidal component, θ: The orientation
of the normal to the parallel stripes of Gabor function, ψ: The
phase offset of the sinusoidal function, θ: The sigma/standard
deviation of the Gaussian envelope, γ: The spatial aspect ratio
and specifies the ellipticity of the support of Gabor function.
3) Nodule Detection: For the nodule detection, Otsu's
method was applied to the image Igabor which is based on
Fig. 2. Proposed CNN model layers. a binary thresholding algorithm that calculates the threshold
values by minimizing the weighted variance of pixels denoted
by σω2 (t). A binarized nodule detected mask was applied to
1) Noise Removal: For the noise removal purpose, Fast the input gray-scale image Igray , and for the stage prediction
Fourier Transform was applied to cancer detected grayscale Hough transform was applied to the nodule detected image
image Icd , which transforms Icd image between spatial and Ind . The whole computation equation of Otsu's thresholding
frequency domains. A low pass filter was applied to the algorithm can be described as,
Fourier transformed image If f t , which is in the frequency
domain to remove high frequencies in If f t image. Then σω2 (t) = ω1 (t)σ12 (t) + ω2 (t)σ22 (t), (6)
inverse Fast Fourier Transformed image Iif f t was passed
through the Gabor filter for clear edge detection. Fourier where, ω1 (t), ω2 (t) are the probabilities of the two classes
Transform for an image size of N × N , divided by a threshold t, which value is within the range from
0 to 255.
N −1 N −1 4) Stage Prediction: Connected component labeling
ak bl
X X
F [Iif f t (k, l)] = f [If f t ](a, b)e−j2π( N + N ) , (1) (CCL) and Hough transform were used for stage prediction
i=0 j=0 [27]. CCL is used in digital image processing to detect a
connected group of pixels in a grayscale image with similar
Inverse Fourier Transform for an image size of N × N ,
pixel intensity values and labeled with a gray level or a
color (color labeling) according to the component. Hough
N −1 N −1
1 X X ak bl transform is a feature extraction method that is mostly used
f [If f t (a, b)] = F [Iif f t ](k, l)ej2π( N + N ) , (2)
N 2 in medical image applications to detect circles and lines in an
k=0 l=0
image. Simple shapes like a circle can be represented by its
where, f [If f t ](a, b) is the image in the spatial domain and center coordinates and radius or a line can be represented by
the exponential term is the basis function corresponding to its slope and intercept. Hough transform will be responsible
each point F [Iif f t ](k, l) in the Fourier space. for processing the dots on the image and calculating the
2) Edge Detection: In the spatial domain, the Gabor values in Hough space based on its parameters. Ind , which is
filter is a Gaussian filter function modulated by a sinusoidal a nodule detected image was passed through Hough transform
function which is used for edge detection. In the process of to detect and calculate the outer diameter of the nodule.
stage prediction, the Gabor filter is used to extract the edges in
Iif f t . Edge detected image Igabor was passed through Otsu's IV. E XPERIMENTAL R ESULTS AND D ISCUSSION
thresholding algorithm to create a nodule detected mask. The The proposed CAD model is capable of detecting cancer
Gabor filter function as given in the following equation has and nodules in the CT image. The model was developed
been applied to each image. and evaluated by an image dataset called VIA (Cornell
Complex: University Vision and Image Analysis Group) and I-ELCAP
(International Early Lung Cancer Action Program), which
2 is provided by the public library [28]. Expert radiologists
x0 + γ 2 y 0 2πx0
g(x, y, λ, θ, ψ, γ, δ) = exp(− 2
)exp(i( + ψ)), confirmed all of the tumor nodules in this dataset, and the
2σ λ
(3) clinical diagnosis results were also provided.
Real:
A. Dataset
x0 + γ 2 y 0
2
2πx0 The dataset used in the training phase is obtained from
g(x, y, λ, θ, ψ, γ, δ) = exp(− )cos(i( + ψ)), the public library provided by VIA and I-ELCAP, which
2σ 2 λ
(4) currently provides the size of 512 × 512 pixels, DICOM
Imaginary: (.dcm) formatted, 12,645 of whole-lungs CT scan image, and

Authorized licensed use limited to: Zhejiang University. Downloaded on September 23,2023 at 02:21:04 UTC from IEEE Xplore. Restrictions apply.
(a) (b) (a) (b) (c)

Fig. 3. Samples of the dataset, (a) normal CT image, and (b) abnormal CT Fig. 5. Image segmentation, (a) original image I, (b) binarized mask Ib ,
image with marked cancer nodule. and (c) segmented image Iseg .

D. Cancer Detection Model


After the image segmentation process, segmented images
were passed through CNN, which is more reliable and
accurate, to train the model for cancer detection. In this study,
(a) (b)
SVM and CNN architecture were tested with the dataset. The
variation in accuracy for incrementing number of layers and
different classifiers is shown in Table IV. When comparing to
other models which we have tested with our dataset, the No.
4 model has the highest accuracy. The output of the neural
(c) (d)
network intermediate layers of the proposed system is shown
in Fig. 6.
Fig. 4. Filtered images, (a) input image, (b) mean filtered image, (c) median The dataset of the lung images taken for experimentation
filtered image, and (d) Gaussian filtered image.
is 100, and from that 60 abnormal (with cancer) images and
40 normal (without cancer) images were randomly selected
it has low noise impression as compared to X-ray and MRI and validated with the CNN model. The confusion matrix for
images. A sample of the dataset is shown in Fig. 3. the CNN model is shown in Table V.
As mentioned in Table V, out of 60 images, 48 images
B. Image Preprocessing of abnormal class are predicted as abnormal, 8 images from
40 normal images are classified as abnormal images, from
The image pre-processing stage starts with image enhance- 60 abnormal images 12 images were classified as normal
ment which improves the interpretability of information in- images, and 32 images from 40 normal images are classified
cluded in the image or provides better input for the automated as normal. After the cancer detection process, cancer detected
image processing model. To enhance the image features, the abnormal images are passed through feature extraction meth-
Mean filter, Median filter, and Gaussian filter were tested with ods, which are Fourier transform and Gabor filter.
the training dataset as shown in Fig. 4.
A mean filter is a non-linear filter that removes noise while E. Noise Removal
keeping edges relatively sharp in the input image I. The Mean Fast Fourier Transform, which is used to decompose an
filter is a simple sliding-window spatial filter that slides the image into its sine and cosine components was tested with
kernel all over the image and replaces the center value in the cancer detected images, and both low pass filter and high pass
kernel with the mean of all the pixel values that cover under filter with Fast Fourier transform were applied for feature
the kernel [29]. The Median filter is a non-linear filter that extraction to detect the nodules (Fig. 7).
replaces the pixel values with the median value available in Low pass filter only allows low frequencies, which means
the 5 × 5 or 3 × 3 pixels around the central pixel value [30]. pixel values that are changing slowly, to pass through the filter
Gaussian filter is a linear filter that is used to blur the image [31]. Similarly, only high frequencies are allowed to pass
or to reduce the noise. through a high pass filter. While the low pass filter removes
After comparing all the above three filters, the Median filter the noisy pixels in the image, the High pass filter captures
removes the noise without blurring the input image, whereas the edges in an image [32].
the mean filter and Gaussian filter remove the noise with the When comparing the low passed filtered image and high
blurred image as the output. passed filtered image, different characteristics were noticed
(Fig. 8). Since low pass filter preserves overall details in the
C. Image Segmentation image, low passed filtered inverse FT image was used to pass
For image segmentation, a binary thresholding algorithm through the Gabor filter to extract the features further.
was used and the threshold value 127-pixel intensity was
selected. Using Canny Edge Detection, which is an OpenCV F. Edge Detection
inbuilt function, the edges of the median filtered image can The Gabor filter was applied to the FFT image, which
be identified. Segmented image results are shown in Fig. 5. is used for edge detection, texture analysis, and feature

Authorized licensed use limited to: Zhejiang University. Downloaded on September 23,2023 at 02:21:04 UTC from IEEE Xplore. Restrictions apply.
TABLE IV. S UMMARY OF T ESTED C ANCER D ETECTION M ODELS
No. Model layer No. of filters/neurons Kernel size Loss function Optimizer No. of epochs Accuracy (%)
Conv2D 64 3×3
Maxpool2D 2×2
1 Flatten Sparse Categorical Cross-entropy adam 10 33.33
Dense 128
Dense 4
Conv2D 32 2×2
Maxpool2D 2×2
Conv2D 32 2×2
2 Maxpool2D 2×2 Categorical Cross-entropy msprop 7 55.11
Flatten
Dense 256
Dense 128
Dense 4
Conv2D 32 4×4
Maxpool2D 2×2
Conv2D 32 4×4
3 Maxpool2D 2×2 Categorical Cross-entropy adadelta 150 61.33
Flatten
Dense 256
Dense 4
Conv2D 32 3×3
Conv2D 32 3×3
Maxpool2D 2×2
Conv2D 64 3×3
4 Conv2D 64 3×3 Sparse Categorical Cross-entropy msprop 10 79.88
Maxpool2D 2×2
Flatten
Dense 512
Dense 4
Support vector machine (SVM)
Sigmoid 0.09
Polynomial 0.53
5
Linear 0.76
Radial basis function (RBF) 0.74

TABLE V. Confusion Matrix for CNN Classification algorithm generates a threshold based on the histogram of
the input image and separates the image histogram into
Actual
Positive Negative two classes (foreground and background) with a result of
Predicted
Positive 48 8 minimization of the weighted variance of the classes (Fig.
Negative 12 32 10).
The Otsu's threshold mask was applied to the input image
and the lung cancer nodules in the abnormal CT image were
extraction by changing the parameters. The optimal parameter detected (Fig. 11).
values were obtained by changing the other parameters as in
Table VI. H. Stage Prediction
The optimal parameter values are, Lambda (λ) = 30; Theta Finally, CCL and Hough transform were applied to the
(θ) = 15◦ , Gamma (γ) = 0.20; Sigma (σ) = 10, Psi (ψ) = 0 nodule detected image to extract the nodules from the image
in Gabor filter. Gabor filter output is shown in Fig. 9. and calculate the outer diameter of the nodule in pixels (Fig.
12).
TABLE VI. G ABOR F ILTER PARAMETER C HANGING S TEPS
Parameter Min value Max value Step
1 pixel = 0.026458333 cm, (7)
Lambda (λ) 0 100 10
Theta (θ) 0◦ 180◦ 15◦ The equation of converting the pixel value into cm is in (7)
Gamma (γ) 0 1 0.25 was used to calculate the outer circle diameter in centimeters.
Sigma (σ) 0 40 5 The algorithms were tested for 40 non-cancer images and
60 cancer images, and the working process for 10 images is
shown in Table VII.
G. Nodule Detection All above-mentioned processes were executed through
Otsu's thresholding method can be used to convert Gabor Python 3.6 in Intel(R) Core (TM) i3 CPU @1.7GHz with
filtered image into a binary nodule detected image. Otsu's 4.0 GB installed memory (RAM) 64-bit windows operating

Authorized licensed use limited to: Zhejiang University. Downloaded on September 23,2023 at 02:21:04 UTC from IEEE Xplore. Restrictions apply.
TABLE VII. F INAL O UTPUT R ESULTS
No. Input Image Input Label Output Image Predicted Label Nodule Size (d) cm Cancer Stage

1 Normal Normal, no cancer detected - -

2 Normal Normal, no cancer detected - -

3 Normal Normal, no cancer detected - -

4 Normal Abnormal, Cancer detected 1.74 I

5 Abnormal Normal, no cancer detected - -

6 Abnormal Abnormal, Cancer detected 0.97 I

7 Abnormal Abnormal, Cancer detected 3.46 II

8 Abnormal Abnormal, Cancer detected 3.76 II

9 Abnormal Abnormal, Cancer detected 5.67 III

10 Abnormal Abnormal, Cancer detected 6.89 III

system and the execution times for the results are shown in Fast Fourier transformation, Gabor filter, Otsus thresholding
Table VIII. algorithm, CCL methods, and Hough transform was proposed
here. The size of the nodule was calculated by measuring the
V. C ONCLUSION
maximum possible outer diameter of the nodule. This system
Due to its severity, early detection of lung cancer is neces- can aid radiologists and doctors in early diagnosis, and this
sary for proper treatment to commence. In this paper, we have implementation can be further improved with higher accuracy
proposed a lung cancer detection method using CNN and in future studies.
lung cancer stage prediction from extracting the features in
the image. To detect the cancer nodules, the CNN algorithm R EFERENCES
was applied to the data set, which is publicly available in VIA
and I-ELCAP. The cancer detection scheme performed at an [1] J. A. Barta, C. A. Powell, and J. P. Wisnivesky, “Global epidemiology
of lung cancer,” Annals of global health, vol. 85, no. 1, 2019.
accuracy of 79.9%. Moreover, a novel methodology for pre- [2] W. H. Organization et al., “Cancer country profile-sri lanka,” in who.
dicting the lung cancer stages using feature extraction such as int, 2014.

Authorized licensed use limited to: Zhejiang University. Downloaded on September 23,2023 at 02:21:04 UTC from IEEE Xplore. Restrictions apply.
(a) (a) (b)

(b)

(c) (d)

Fig. 8. Low passed and high passed filtered image, (a) Low pass filter, (b)
Low passed filtered inverse FT image Iif f t , (c) High pass filter, and (d)
(c)
High passed filtered inverse FT image.

(d)

(a) (b)

Fig. 9. Gabor filter enhancement technique, (a) Fourier transformed image


If f t , and (b) Gabor filtered image Igabor .

(e)

(a) (b)
(f)

Fig. 6. Intermediate layers of the proposed neural network model, (a) first
convolution layer with 32 filters, (b) second convolution layer with 32 filters,
(c) first max-pooling layer, (d) second convolution layer with 64 filters, (e)
second convolution layer with 32 filters, and (f) second max-pooling layer.

(c) (d)

Fig. 10. Otsu's thresholding algorithm, (a) Fourier transformed image If f t ,


(b) Histogram of FT image, (c) Otsu's threshold mask, and (d) Histogram
of threshold image.

(a) (b)

Fig. 7. Fourier transformation of an image, (a) Input image I, and (b) (a) (b)
Fourier transformed image If f t .
Fig. 11. Nodule detected image, (a) Input image I, and (b) Nodule detected
image Icd .

Authorized licensed use limited to: Zhejiang University. Downloaded on September 23,2023 at 02:21:04 UTC from IEEE Xplore. Restrictions apply.
Systems (ICECS 2012). IEEE, 2012, pp. 773–776.
[11] P. Naresh, D. R. Shettar et al., “Early detection of lung cancer using
neural network techniques,” Int Journal of Engineering, vol. 4, pp.
78–83, 2014.
[12] S. G. Armato III, L. Hadjiiski, G. D. Tourassi, K. Drukker, M. L.
Giger, F. Li, G. Redmond, K. Farahani, J. S. Kirby, and L. P. Clarke,
(a) (b) (c) “Spie-aapm-nci lung nodule classification challenge dataset. the cancer
imaging archive,” 2015.
[13] P. Li, S. Wang, T. Li, J. Lu, Y. HuangFu, and D. Wang, “A large-scale
ct and pet/ct dataset for lung cancer diagnosis.”
[14] P. Kalendralis, Z. Shi, A. Traverso, A. Choudhury, M. Sloep, I. Zho-
vannik, M. P. Starmans, D. Grittner, P. Feltens, R. Monshouwer
et al., “Fair-compliant clinical, radiomics and dicom metadata of rider,
interobserver, lung1 and head-neck1 tcia collections,” Medical Physics,
(d) (e) vol. 47, no. 11, pp. 5931–5940, 2020.
[15] A. Madabhushi and M. Rusu, “Fused radiology-pathology lung
Fig. 12. Computed outer circle of the nodule, (a) Input image, (b) Otsu's dataset,” The Cancer Imaging Archive.
threshold image, (c) CCL image, (d) Nodule detected image, and (e) Outer [16] J. Yang, “Data from lung ct segmentation challenge,” The Cancer
circle of the nodule. Imaging Archive.
[17] A. Asuntha and A. Srinivasan, “Deep learning for lung cancer detection
and classification,” Multimedia Tools and Applications, vol. 79, no. 11,
TABLE VIII. E XECUTION T IME FOR E ACH P ROCESS pp. 7731–7762, 2020.
[18] M. Toğaçar, B. Ergen, and Z. Cömert, “Detection of lung cancer
Execution time (s)
on chest ct images using minimum redundancy maximum relevance
Preprocessing and CNN model

feature selection method with convolutional neural networks,” Biocy-


Fast Fourier transform

bernetics and Biomedical Engineering, vol. 40, no. 1, pp. 23–39, 2020.
Total execution time
Otsus thresholding

[19] K. Clark, B. Vendt, K. Smith, J. Freymann, J. Kirby, P. Koppel,


Hough transform
Gabor filter

S. Moore, S. Phillips, D. Maffitt, M. Pringle et al., “The cancer


imaging archive (tcia): maintaining and operating a public information
CCL
No.

repository,” Journal of digital imaging, vol. 26, no. 6, pp. 1045–1057,


2013.
[20] Q. Li, W. Cai, X. Wang, Y. Zhou, D. D. Feng, and M. Chen,
“Medical image classification with convolutional neural network,” in
2014 13th international conference on control automation robotics &
vision (ICARCV). IEEE, 2014, pp. 844–848.
1 2.03 - - - - - 2.03 [21] B. Xu, “Identifying fabric structures with fast fourier transform tech-
2 2.60 - - - - - 2.60 niques,” Textile Research Journal, vol. 66, no. 8, pp. 496–506, 1996.
3 2.50 - - - - - 2.50 [22] S. Avinash, K. Manjunath, and S. S. Kumar, “An improved image
4 2.64 0.79 0.62 0.91 0.79 1.03 6.78 processing analysis for the detection of lung cancer using gabor
5 2.85 - - - - - 2.85 filters and watershed segmentation technique,” in 2016 International
6 2.28 0.79 0.67 1.01 0.79 0.98 6.52 Conference on Inventive Computation Technologies (ICICT), vol. 3.
7 2.35 0.76 0.67 1.09 0.77 0.93 6.57 IEEE, 2016, pp. 1–6.
8 2.32 0.78 0.59 0.91 0.79 1.09 6.48 [23] T. Y. Goh, S. N. Basah, H. Yazid, M. J. A. Safar, and F. S. A.
9 2.02 0.85 0.55 1.02 0.76 1.02 6.22 Saad, “Performance analysis of image thresholding: Otsu technique,”
10 2.09 0.84 0.63 0.91 0.72 1.01 6.20 Measurement, vol. 114, pp. 298–307, 2018.
[24] H. Yuen, J. Princen, J. Illingworth, and J. Kittler, “Comparative study
of hough transform methods for circle finding,” Image and vision
computing, vol. 8, no. 1, pp. 71–77, 1990.
[25] D. Bradley and G. Roth, “Adaptive thresholding using the integral
[3] R. Jha, A. Weerakoon, K. Karki, S. Shrestha, and P. Gamage, “Lung image,” Journal of graphics tools, vol. 12, no. 2, pp. 13–21, 2007.
cancer and smoking in asia,” SAARC Journal of Tuberculosis, Lung [26] A. Polesel, G. Ramponi, and V. J. Mathews, “Image enhancement via
Diseases and HIV/AIDS, vol. 5, no. 1, pp. 33–38, 2008. adaptive unsharp masking,” IEEE transactions on image processing,
[4] D. R. Varma, “Managing dicom images: Tips and tricks for the vol. 9, no. 3, pp. 505–510, 2000.
radiologist,” The Indian journal of radiology & imaging, vol. 22, no. 1, [27] L. He, Y. Chao, K. Suzuki, and K. Wu, “Fast connected-component
p. 4, 2012. labeling,” Pattern recognition, vol. 42, no. 9, pp. 1977–1987, 2009.
[5] B. Halalli and A. Makandar, “Computer aided diagnosis-medical image [28] “Via,” accessed Jun. 03, 2021). [Online]. Available:
analysis techniques,” Breast Imaging, vol. 85, 2018. https://www.via.cornell.edu/research/
[6] F. C. Detterbeck, D. J. Boffa, and L. T. Tanoue, “The new lung cancer [29] X. Zhang and Y. Xiong, “Impulse noise removal using directional
staging system,” Chest, vol. 136, no. 1, pp. 260–271, 2009. difference based noise detector and adaptive weighted mean filter,”
[7] A. Kanakatte, N. Mani, B. Srinivasan, and J. Gubbi, “Pulmonary tumor IEEE Signal processing letters, vol. 16, no. 4, pp. 295–298, 2009.
volume detection from positron emission tomography images,” in 2008 [30] T. Sun and Y. Neuvo, “Detail-preserving median based filters in image
International Conference on BioMedical Engineering and Informatics, processing,” Pattern recognition letters, vol. 15, no. 4, pp. 341–347,
vol. 2. IEEE, 2008, pp. 213–217. 1994.
[8] Y. Lee, T. Hara, H. Fujita, S. Itoh, and T. Ishigaki, “Automated [31] O. K. Khorsheed, “Produce low-pass and high-pass image filter in
detection of pulmonary nodules in helical ct images based on an java,” International Journal of Advances in Engineering & Technology,
improved template-matching technique,” IEEE Transactions on medical vol. 7, no. 3, p. 712, 2014.
imaging, vol. 20, no. 7, pp. 595–604, 2001. [32] X. Chen, J. Yang, Q. Wu, J. Zhao, and X. He, “Directional high-pass
[9] S. H. Hawkins, J. N. Korecki, Y. Balagurunathan, Y. Gu, V. Kumar, filter for blurry image analysis,” Signal Processing: Image Communi-
S. Basu, L. O. Hall, D. B. Goldgof, R. A. Gatenby, and R. J. Gillies, cation, vol. 27, no. 7, pp. 760–771, 2012.
“Predicting outcomes of nonsmall cell lung cancer using ct image
features,” IEEE access, vol. 2, pp. 1418–1426, 2014.
[10] F. Taher, N. Werghi, and H. Al-Ahmad, “Bayesian classification and
artificial neural network methods for lung cancer early diagnosis,” in
2012 19th IEEE International Conference on Electronics, Circuits, and

Authorized licensed use limited to: Zhejiang University. Downloaded on September 23,2023 at 02:21:04 UTC from IEEE Xplore. Restrictions apply.

You might also like