Download as pdf or txt
Download as pdf or txt
You are on page 1of 9

Journal of Ambient Intelligence and Humanized Computing (2023) 14:4675–4683

https://doi.org/10.1007/s12652-022-04368-w

ORIGINAL RESEARCH

A new framework for multi‑scale CNN‑based malignancy classification


of pulmonary lung nodules
Sakshiwala1 · Maheshwari Prasad Singh1

Received: 2 November 2021 / Accepted: 30 July 2022 / Published online: 18 September 2022
© The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2022

Abstract
Lung cancer has the highest mortality rate among all types of cancers. Early detection of lung cancer may improve survival
rates. The two categories of pulmonary lung nodules have high visual similarities. So, distinguishing them is a challenging
task for radiologists. The main purpose of this work is to use convolutional neural network to perform binary classification
of pulmonary nodules in CT images. This paper proposes a new multi-scale (64 × 64, 32 × 32 and 16 × 16) convolutional
neural network architecture for benign and malignant nodules classification. In addition, transfer learning method is used
to initialize the weights of multi-scale architecture. Experimental results on the dataset LIDC-IDRI demonstrate that the
proposed method achieved accuracy of 93.88%, sensitivity of 93.36% and specificity of 93.26% on nodule malignancy
classification. The proposed method also outperforms the other state-of-the-art methods explicitly designed for malignancy
classification of pulmonary lung nodules.

Keywords Transfer learning · Benign · Malignant · LIDC-IDRI · CT scan · Lung cancer

1 Introduction (CAD) systems can be used to reduce these variations. CAD


systems are able to provide the accurate, quick, and cost-
Lung cancer is the most lethal form of cancerous disease. effective solutions and thus have a clear clinical importance.
It is uncontrolled growth of cells in the lung tissues. Every In recent years, deep neural network based methods have
year it affects 12.9% of the world’s population. In less devel- shown good performance in image based medical diagno-
oped countries there is 58% of lung cancer cases (Abid et al. sis, and thus has an important role in the design of CAD
2021). Typically symptoms of this disease appear only when systems.
it is at an advanced stage. So in order to avoid spread in Doctors predict nodule malignancy risk on the basis of
the body (Chen et al. 2020; Attique et al. 2020; Zhao et al. shape and density of the nodules and the other pathology
2018a), lung cancer should be treated at an early stage. information. Qualitatively, compared to the benign nodules,
Lung cancer diagnosis by analysing CT scans is clini- the malignant often has irregular shape, higher volumes, lob-
cally intensive task for radiologists. It takes ten minutes for ulated and spiculated margin. Some images of pulmonary
the radiologist to give a thorough inspection of a CT scan. lung nodules are shown in Fig. 1. To overcome the prob-
Moreover, there is considerable variation among radiologists lems mentioned above, this paper proposes a new frame-
in identifying nodules. However, this process is influenced work using multi-scale convolutional neural network (CNN)
by the emotion and experience of the respective doctors. architecture with transfer learning approach. The proposed
The overall error rate is reported to be 7.6% for every 53 method adopts a transfer learning approach to preserve the
patients (Venugopal et al. 2020). Computer-aided diagnosis learned features using multiple filters with different sizes.
Contributions of this paper can be summarized as follows:
* Sakshiwala
sakshiwala.phd19.cs@nitp.ac.in – A new framework for multi-scale CNN-based architec-
Maheshwari Prasad Singh ture is proposed to classify the lung nodules into benign
mps@nitp.ac.in and malignant classes. A 64 × 64 pixels input based
model is first trained.
1
National Institute of Technology Patna, Patna, Bihar 800005,
India

13
Vol.:(0123456789)
4676 Sakshiwala, M. P. Singh

and identity blocks of the encoder are passed to each respec-


tive layers of classification network.
Yu et al. (2018) propose a density-based spatial clustering
of applications with noise algorithm. The advantage is that
one can choose the best scale combination from multi-scale
(a) Benign nodules
according to the need. The method is not able to detect all
types of nodules.
In Hussein et al. (2019), the model uses weights pre-
trained on sports dataset. It is multi-task learning on visual
attributes. K-means clustering and proportion-SVM. The
(b) Malignant nodules unsupervised learning methods like GANs and autoencod-
ers are more promising. However, unsupervised learning
Fig. 1  Pulmonary lung nodule sample images schemes require manual feature extractions, so it is complex
to find the optimal feature combinations.
The most popular architecture of DNN is CNN. CNN
– Then the transfer learning method is used to train the 32 and many of its variants are based on supervised learning
× 32 pixels and 16 × 16 pixels input based models. methods. From raw input images, CNN is capable to extract
high-level features without any manual intervention. Thus, it
The paper is organized as follows: Sect. 2 describes the is expected to perform better in classification of pulmonary
Related Work. Section 3 describes the proposed methodol- nodules.
ogy. Section 4 demonstrates experimental results and discus- Pritam Mukherjee et al. (2020) develop LungNet, a shal-
sion. Section 5 concludes the paper. low convolutional neural network. It has two CNNs, the
survival prediction and the malignancy prediction network.
The weights of the survival prediction network is shared by
2 Related work transfer learning with the malignancy prediction network.
They demonstrated the use of transfer learning. However, the
Various studies are present in the literature that have the aim method extract features by single-scale filter. However, due
of detection and classification of lung cancer nodules using to variation in nodule sizes, analyzing nodule images using
both the deep learning and conventional machine learning multi-scale filters could generate more effective nodule fea-
approaches. Early developed CAD systems use low level fea- tures. It has better accuracy compared to the conventional
tures of medical imaging, which represent the global struc- feature extraction methods. In Zhao et al. (2018b) combines
tures in the images such as shape and size. These methods the benefits of LeNet layer settings and AlexNet parameter
are useful to classify pulmonary nodules because it has dif- settings. Since LeNet and AlexNet are originally designed
ferent sizes and shapes. for coloured images, this affects the performance because
However, visual representations may be different for dif- the medical images are gray coloured images.
ferent types of the nodules and needs to be differentiated Zhang et al. (2019) develop a meta-learning method,
with both global and local structures. The high-level fea- where the output of the primary learners is input dataset to
tures of nodules may be extracted by unsupervised feature the secondary learners. The eight variants of CNN predict
learning methods. It may achieve better classification per- the malignancy value for each patch. The predicted values
formance compared to the low-level based feature extractors. for each patch are then used as a new dataset for the ten
Kuang et al. (2020) train the network to learn features of ensemble learners. The computational cost involved in this
anomalous benign images. And tests the network using non- method might be very high.
anomalous images. The method requires manual estimation To fulfill the need of an accurate and robust pulmonary
of the malignancy scoring threshold. Xie et al. (2019) sug- nodule classification, in this paper, a novel convolutional
gests a multi-view knowledge-based semi-supervised adver- neural network baesd architecture is proposed. The proposed
sarial classification model. It is the extension of the work method adopts a multi-scale feature extraction scheme while
proposed in Xie et al. (2018a). The network uses semi-super- using the benefits of transfer learning approach. The multi-
vised reconstruction network trained on labelled and unla- scale feature have better performance along with multiple
belled data. The features extracted from each convolution learners. The transfer learning scheme shows promising
results for the nodules classification.

13
A new framework for multi‑scale CNN‑based malignancy classification of pulmonary lung nodules 4677

3 Methods reading the corresponding xml file. The four radiologists


have independently annotated each nodule, and has given
This section describes the database and preprocessing meth- malignancy score that range from 0 to 5. The mean value of
ods used in this paper. The method and the architectural these scores is taken as the final malignant score value for
details of the proposed model are discussed. that particular nodule. This is phrased as mean malignancy
label of a nodule. The nodules with mean malignancy label
3.1 Dataset description and preperation equal to 3 were removed from this study. The nodules are
further categorized using mean malignancy label. Finally, a
The Lung Image Database Consortium image collection total of 829 malignant nodules and 1262 benign nodules are
(LIDC-IDRI) (Samuel et al. 2004), initiated by the National extracted. Figure 2 illustrates the described preprocessing.
Cancer Institute, is used to train and test the proposed
method. Total of 1012 chest CT scans were considered for 3.2 Proposed model
this study. Each CT scan has an xml annotation file attached,
which contains information of nodules marked by four expe- Below mentioned subsections provide the description of the
rienced annotated radiologists. Depending on the diameter various layers used in the proposed framework and the pro-
of the nodules, the radiologist categorised the nodules into posed multi-scale architecture.
three categories: nodules > 3 mm, nodules < 3 mm and non- 2D convolution layer: Convolution layer performs linear
nodule >= 3 mm. convolution operation on the input. The convolution opera-
The original CT images of 512 × 512 pixels in .dcm tion involves multiplication of an array of input data with
format is converted into .png images for faster training. a filter of weights. Mathematically, convolution is a feature
Since the nodules size is very small compared to the whole derived from two given functions by integration which
image size, it is required to extract the particular region of expresses how one function transforms the shape of other.
interest of nodules. According to the requirement, first, we Figure 3 shows a 2D convolution operation. In order to get
segment the lung parenchyma from CT images to remove multiple convolved images, various kind of feature detectors
the irrelevant information. Then using threshold probabil- are applied on input image to extract different features.
ity map algorithm, the 64 × 64 pixels rectangular region Maxpooling layer: Max pooling operation is applied on
is intercepted with the pulmonary nodules. The location each patch of each feature map seperately. It pools out the
information of nodules (Kuang et al. 2020) is obtained by maximum values and form a set of same number of feature
maps. While it reduces the size of each feature map. Pooling
helps to downsample the feature maps. It is important that
it preserves the original feature and still able to reduce the
spatial size and thus computation in the network. An sample
of max pooling operation is shown in Fig. 4. The benefit of

Fig. 3  Convolution layer operation

Fig. 2  Illustrations of preprocessing Fig. 4  Maxpooling operation

13
4678 Sakshiwala, M. P. Singh

maxpooling is that it pools out that feature from the feature LeakyReLU turns all negative input values to a value near
map which is most present. to zero. The idea is to let some gradient flow during training.
LeakyReLU operation: In order to get the desired While all positive values remain same.
results, activation functions is used to make decision on Flatten and the dense layer: Flatten layer translates the
whether to activate or deactivate a neuron. It helps model to multi-dimensional tensor into a single long feature vector.
learn non-linear transformations. Dense layers are the fully connected layers. Unlike convolu-
tion layers where a neuron is connected to local area of input
neurons, in the dense layer each neuron has a path to every
other neuron to next layer. Figure 5 illustrates the flatten and
dense layer.

3.2.1 Architecture details of the proposed framework

The framework of the proposed model is shown in Fig. 6b.


The proposed method has three models that perform multi-
scale feature extraction. The first model takes input of size
64 × 64 pixels. The second model takes input of size 32 ×
32 pixels and third model takes 16 × 16 pixels. Each of the
model has the same architectural details. The first layer uses
filters of size 5 × 5 to perform the convolution operation on
the input image and generates 64 output feature maps. The
second, third and fourth convolution layer uses 3 × 3, 5 ×
Fig. 5  Flatten and dense layer 5 and 3 × 3 sized filters to perform convolution operations

(a) training (b) testing

Fig. 6  Proposed model

13
A new framework for multi‑scale CNN‑based malignancy classification of pulmonary lung nodules 4679

respectively. The LeakyReLU activation function is used Such as rotation by 90°, 180°and 270° angle. The nodule
after every convolutional layer. It helps the model to con- images were cropped and flipped. The nodules are aug-
verge faster and is easy to implement. mented by translating the nodules images by x axis and
Then, this is followed by a max-pooling layer. This layer y axis ± 2 pixels. Data augmentation techniques helps to
reduces the spatial resolution and helps in learning the more generalize the model. Adam optimizer with learning rate of
complex features from the input. Further two dense layers 0.001 is used.
are used. The flattened output feature maps are then passed All experiments in this paper were performed on Python
to a dense layer of 512 neurons. After this, LeakyReLU programming language using Keras library. The net-
activation function is used. The last fully connected layer works were trained on 32 GB RAM with a single NVIDIA
has two neurons which classifies the nodules as benign and QUADRO M2000, Windows 10 operating system. Results
malignant. The outputs from the last fully-connected layer are given below with discussions.
is normalized using softmax. The fully-connected layers
often lead to overfitting. The problem of overfitting can be 4.2 Performance metrics
resolved using dropout. A dropout of 0.5 is used to avoid
the overfitting. A sparse categorical crossentropy loss (L) Different evaluation parameters, including accuracy, preci-
function is used. It is defined as: sion, recall (sensitivity or True positive rate), F1-score, spec-
2
ificity, false positive rate are used to evaluate the proposed
∑ framework. Some parameters depend on the outcome of the
L=- log(p(i))t(i), (1)
x=1
confusion matrix true positive (TP), false negative (FN), true
negative (TN) and false positive (FP). TP is the number of
where p(i) is the predicted probability, t(i) is the true prob- correctly predicted malignant nodules. TN is the number
ability, ‘x’ is the class label and ‘i’ is the sample training of correctly predicted benign nodules. FP is the number of
data. Smaller value of loss function implies that the pre- actual benign nodules that are predicted malignant. FN is the
dicted probability is closer to the true probability. actual malignant nodules that are predicted benign.
The statistical parameters are expressed as follows:
3.2.2 Multi‑scale feature extraction and transfer learning TP + TN
Accuracy = (2)
TP + FP + TN + FN
The purpose of multi-scale feature extraction is to learn dif-
ferent patterns in images using different filter sizes. Differ- TP
ent filter sizes can focus on the context information of the Precision = (3)
TP + FP
candidate nodules in different scale images and the features
extracted from the different streams can also complement TP
each other. Therefore, more effective features can be learnt Recall = (4)
TP + FN
through different scale images. The method uses transfer
learning concept to transfer the weights learned by convo- TN
lutional layer of 64 × 64 pixels input based model to the Specificity = (5)
TN + FP
respective layers of 32 × 32 pixels input based model. The
layers are then trained. The weights after training of 32 ×
FP
32 pixels input based model is transferred to the 16 × 16 FPR = . (6)
TN + FP
pixels input based model. Figure 6b shows the described
framework.
4.3 Tuning of hyperparameters

4 Results and discussion A series of experiments is carried out to justify the choices:
the number of convolutional layers, maxpooling layers, fully
4.1 Experimental environment connected layers, size and number of filters, increasing or
decreasing the batch size, initializing convolutional layer
This work uses dataset having 2091 images. To increase the weights. Table 1 shows accuracies obtained by different sets
randomness data augmentation techniques are employed. of hyperparameters.

13
4680 Sakshiwala, M. P. Singh

Table 1  Tuning of hyperparameters of the proposed model


Input scale Conv layers Fully con- Filter size Dropout ratio Batch size Weight initialization Accuracy
nected layer

64 × 64 4 2 5 × 5, 3 × 3 0.5 96 Random 92.27


64 × 64 4 2 5 × 5, 3 × 3 0.5 128 Random 92.55
64 × 64 4 3 5 × 5, 3 × 3 0.5 224 Random 92.65
64 × 64 3 2 5 × 5, 3 × 3 0.5 224 Random 93.21
64 × 64 4 2 5 × 5, 3 × 3 0.5 224 Random 93.88

(a) Conv1

Fig. 7  Illustration of filters in Conv1 layer of 64 × 64 input size


model

4.4 Effectiveness of multi‑scale features


and usefulness of transfer learning approach

Figure 7 illustrates 64 filters used in the first convolution


layer of a 64 × 64 pixels input based model. These filters
slide over the input image to learn different portions of an
input image. Figures 8a, b, 9a, b represents the output feature
(b) Conv2
maps obtained from first and second convolution layers. As
64 filters are used in first convolution layer, so 64 output fea-
Fig. 8  Representing the hidden layer output activation maps of a 64 ×
ture maps are generated. While 32 filters are used in second 64 benign nodule image
convolution layer, so 32 output feature maps are generated.
Figure 10a–c show the training and validation accuracy
graph of the model on respective input sizes. From the the training, validation and test datasets. In order to avoid
graph, it can be concluded that the model is best trained. the overfitting, patience parameter is used as early stopping
Table 2 shows the classification accuracies of the different criteria during training.
input sizes of the model. The results shown in Table 2 are In Table 3, the accuracy of the model is shown under
when the convolutional layers of 32 × 32 model are initial- the following conditions: by using the transfer learning
ized by the weights of 64 × 64 pixels input based model. approach, the weights learned by filters of first convolutional
The convolutional layers of 16 × 16 model are initialized layer of input size 64 × 64 model is transferred to the 32 ×
by the weights of 32 × 32 pixels input based model. Table 2 32 model. The weights learned by filters of first and second
shows the classification accuracy of the proposed method on

13
A new framework for multi‑scale CNN‑based malignancy classification of pulmonary lung nodules 4681

Table 2  Classification Input size Train Val Test


accuracies of the proposed
model with transfer learning, 64 × 64 0.95 0.94 0.93
non-trainable layers in 16 × 16
32 × 32 0.93 0.93 0.93
input based model is conv1
16 × 16 0.92 0.92 0.92

Table 3  Classification Input size Train Val Test


accuracies of the proposed
Model with transfer learning, 64 × 64 0.94 0.92 0.93
non-trainable layers in 32 × 32
32 × 32 0.93 0.92 0.94
is conv1 and 16 × 16 are conv1
and conv2 16 × 16 0.91 0.91 0.92
(a) Conv1

4.5 Effectiveness of the proposed model

A contingency table called confusion matrix shown in


Fig. 11 depicts how many nodules are correctly classified
and how many were not correctly predicted. 774 nodules
were actually malignant and are correctly predicted as
malignant. 1177 nodule images are actually benign and
are correctly predicted as benign. While 55 nodules were
actually malignant but are wrongly predicted as benign. 85
benign nodules are predicted as malignant.
(b) Conv2
The confusion matrix enteries (TP, TN, FP, FN) are not
meaningful independently, so it is interpreted relative to each
Fig. 9  Representing the hidden layer output activation maps of a 64 ×
other. So we plot ROC curve in Fig. 12. It is the plot between
64 malignant nodule image sensitivity (true positive rate) and 1-specificity (false posi-
tive rate). It shows that the model achieves area under curve
(AUC) of 93.315%. Table 4 shows the precision, recall and
convolution layers of 32 × 32 model is transferred to respec- F1-score of respective classification categories.
tive layers of 16 × 16 model.
4.6 Comparison with previous work

To illustrate the effectiveness of the proposed method,


Table 5 provides the comparison of the proposed method

(a) 64*64 input model (b) 32*32 input model (c) 16*16 input model

Fig. 10  Training graph of model

13
4682 Sakshiwala, M. P. Singh

with other deep learning based methods. All the models


use LIDC-IDRI database for experiments. The number of
data used in Xie et al. (2018b) and Ren et al. (2020) is
relatively small, so the final accuracy rate is lower than
the other studies. Xie et al. (2019) is a semi-supervised
learning approach based on both labeled and unlabeled
data. In addition, it can be seen from Table 4 that the
proposed model has a significant improvement in clas-
sification performance.

5 Conclusion
Fig. 11  Confusion matrix (label ‘0’: benign, label ‘1’: malignant)
In this paper, a new framework of multi-scale CNN-
based architecture is proposed for malignancy classifi-
cation of lung nodules on chest CT scans. The proposed
method adopts a transfer learning based multi-scale fea-
ture extraction method. In this method first a size of 64
× 64 pixels input images is fed as an input to the network
and trained. Then the learned weights are transferred to
the 32 × 32 pixels input based model of same architec-
ture and is then trained. The weights learned by convolu-
tional layers of 32 × 32 pixels input size based model is
initialized to the respective layers of 16 × 16 pixels input
based models and is then trained. Finally, the model
is evaluated on LIDC-IDRI dataset. The experimental
results showed that the proposed model is efficient for
Fig. 12  ROC curve malignancy classification of lung nodules in chest CT
scans.
In this work only 2D image of a single nodule is used
for cancer prediction. While in reality, the oncologists
Table 4  Performance of the proposed model
consider 3D view of CT scans. The current study does not
consider scores of attributes related to malignancy. For
Category Precision Recall F1-score the future work, in order to avoid human cropped nodule
Benign 0.96 0.93 0.94 regions, deep learning based malignancy classification
Malignant 0.90 0.93 0.92 method along with the method of nodule detection may
be added.

Table 5  Comparison with previous models


References Method Database (number) Accuracy Sensitivity Specificity AUC​

Xie et al. (2018b) Fuse-TSD LIDC-IDRI (1972) 89.53 84.19 92.02 96.65
Ren et al. (2020) Manifold regularized CNN LIDC-IDRI (1226) 90 81 95 -
Xie et al. (2019) Semi-supervised adversarial classification LIDC-IDRI (2557) 92.53 84.94 96.28 95.81
Proposed method Multi-scale CNN + transfer learning LIDC-IDRI (2091) 93.88 93.36 93.26 93.31

13
A new framework for multi‑scale CNN‑based malignancy classification of pulmonary lung nodules 4683

Data availibility statement The datasets used in the current study are deep learning era: novel supervised and unsupervised learning
available in the below https://​wiki.​cance​rimag​ingar​chive.​net/​displ​ay/​ approaches. IEEE Trans Med Imaging 38(8):1777–1787
Public/​LIDC-​IDRI. Venugopal VK, Vaidhya K, Murugavel M, Chunduru A, Mahajan V,
Vaidya S, Mahra D, Rangasai A, Mahajan H (2020) Unboxing ai-
radiological insights into a deep neural network for lung nodule
Declarations characterization. Acad Radiol 27(1):88–95
Xie Y, Xia Y, Zhang J, Song Y, Feng D, Fulham M, Cai W (2018)
Conflict of interest The authors have no conflict of interest. Knowledge-based collaborative deep learning for benign-malig-
nant lung nodule classification on chest ct. IEEE Trans Med Imag-
ing 38(4):991–1004
Xie Y, Zhang J, Xia Y, Fulham M, Zhang Y (2018) Fusing texture,
shape and deep model-learned information at decision level for
References automated classification of lung nodules on chest ct. Inf Fusion
42(102–110):7. https://​doi.​org/​10.​1016/j.​inffus.​2017.​10.​005
Abid MMN, Zia T, Ghafoor M, Windridge D (2021) Multi-view con- Yan K, Tian L, Xueqiao P, Elvis SG, Qiao L, Junyi Z (2020) Unsuper-
volutional recurrent neural networks for lung cancer nodule iden- vised multi-discriminator generative adversarial network for lung
tification. Neurocomputing 453:299–311 nodule malignancy classification. IEEE Access 8:77725–77734.
Armato SG III, McLennan G, McNitt-Gray MF, Meyer CR, Yankel- https://​doi.​org/​10.​1109/​ACCESS.​2020.​29879​61
evitz D, Aberle DR, Henschke CI, Hoffman EA, Kazerooni EA, Yu G, Xiaoqi L, Yang L, Zhang B, Dahua Yu, Zhao Y, Gao L, Liang
MacMahon H et al (2004) Lung image database consortium: W, Zhou T (2018) Automatic lung nodule detection using a 3d
developing a resource for the medical imaging research commu- deep convolutional neural network combined with a multi-scale
nity. Radiology 232(3):739–748 prediction strategy in chest cts. Comput Biol Med 103:220–231
Attique Khan M, Sadia R, Asifa K, Imran SM, Nazeer M, Hussain SJ, Yutong X, Jianpeng Z, Yong X (2019) Semi-supervised adversarial
Yu-Dong Z, Chandra SS (2020) Lungs cancer classification from model for benign-malignant lung nodule classification on chest
ct images: An integrated design of contrast based classical features ct. Med Image Anal 57(10):237–248. https://​doi.​org/​10.​1016/j.​
fusion and selection. Pattern Recogn Lett 129:77–85 media.​2019.​07.​004
Baihua Z, Shouliang Q, Patrice M, Chen L, Fan Y, Yu-Dong Y, Wei Zhao W, Yang J, Sun Y, Li C, Wu W, Jin L, Yang Z, Ni B, Gao P,
Q (2019) Ensemble learners of multiple deep cnns for pulmonary Wang P, Hua Y, Li M (2018) 3d deep learning from ct scans pre-
nodules classification using ct images. IEEE Access 7(8):110358– dicts tumor invasiveness of subcentimeter pulmonary adenocar-
110371. https://​doi.​org/​10.​1109/​access.​2019.​29336​70 cinomas. Cancer Res 78(12):6881–6889. https://​doi.​org/​10.​1158/​
Chen Y, Wang Y, Fei H, Wang D (2020) A lung dense deep convolution 0008-​5472.​CAN-​18-​0696
neural network for robust lung parenchyma segmentation. IEEE Zhao X, Liu L, Qi S, Teng Y, Li J, Qian W (2018) Agile convolu-
Access 8:93527–93547 tional neural network for pulmonary nodule classification using ct
Pritam Mukherjee M, Zhou EL, Schicht A, Balagurunathan Y, Napel images. Int J Comput Assist Radiol Surg 13(4):585–595
S, Gillies R, Wong S, Thieme A, Leung A et al (2020) A shallow
convolutional neural network predicts prognosis of lung cancer Publisher's Note Springer Nature remains neutral with regard to
patients in multi-institutional computed tomography image data- jurisdictional claims in published maps and institutional affiliations.
sets. Nat Mach Intell 2(5):274–282
Ren Y, Tsai M-Y, Chen L, Wang J, Li S, Liu Y, Jia X, Shen C (2020) A Springer Nature or its licensor holds exclusive rights to this article under
manifold learning regularization approach to enhance 3d ct image- a publishing agreement with the author(s) or other rightsholder(s);
based lung nodule classification. Int J Comput Assist Radiol Surg author self-archiving of the accepted manuscript version of this article
15(2):287–295 is solely governed by the terms of such publishing agreement and
Sarfaraz H, Pujan K, Bolan Candice W, Wallace Michael B, Ulas applicable law.
B (2019) Lung and pancreatic tumor characterization in the

13

You might also like