Professional Documents
Culture Documents
First Progress Report
First Progress Report
1. Title page
2. Abstract (also contains problem statement)
3. Introduction
4. Literature survey
5. Research gaps
6. Objectives
7. Methodology of study
8. Work left to do.
9. Limitations
10. References
Abstract
In this project, we will implement a deep learning pipeline for disease classification using a deep
learning model. We also aim to present a comparison of various deep learning models for this
application in cohesion with our data preprocessing pipeline. We want to examine if smaller
models can perform as well as some deep CNNs because of data preprocessing. For this we are
using OpenCV and PyTorch.
Introduction
Every year millions of adults are diagnosed with pulmonary pathologies like: pulmonary
atelectasis, cardiomegaly, pulmonary effusion, infiltration, pulmonary mass, nodule, pneumonia,
pneumothorax [2,3], etc. According to the Indian Council of Medical Research, in 2019 alone,
about 1.6 million people died due to COPD (Chronic obstructive pulmonary disease) [24]. Also,
the COVID-19 pandemic truly took our healthcare infrastructures abase, causing millions of
people to die due to breathing difficulty caused by pulmonary fibrosis (a symptom of Covid-19).
CXRs (Chest X-Rays) are currently one of the best available methods for diagnosing pulmonary
pathologies [3], playing a crucial role in clinical healthcare and epidemiological studies. But,
diagnosing the X-ray images with correct pathology is an involved task and relies heavily upon
the availability of expert radiologists.
In this project, we propose a novel deep learning based pipeline for classifying aforementioned
pulmonary pathologies from CXRs. While other researchers have used deep learning for this
application [7], we are using a combination of preprocessing techniques such as Bone
Suppression [13] and HE (Histogram Equalization)[25] to enhance our model’s capability.
Our project aims to help our healthcare workers provide quality diagnosis even to the most
remote places quickly and cheaply.
Literature Survey
The application of Deep Convolutional Neural Networks for disease classification using CXRs
really became popular with the RSNA Pneumonia Detection Challenge [1] on Kaggle after the
Stanford team consisting of Andrew Ng et al. submitted their CheXnet model [7], which was
based on a Dense CNN. The CheXnet team used Chest X-ray 14 dataset which contained about
112000 X-rays with annotations of up to 14 different thoracic pathologies, released by Wang et
al. in 2017 [5,6]. Concomitantly, they published a research paper on CXRs disease classification
in which they shared their model’s performance. CheXnet achieved F1 scores of around 0.435
which was significantly higher than compared to the F1 score, 0.387 of Stanford radiologists.
Their paper proved the obtainability of Deep Learning techniques for chest X-ray disease
classification.
Then later in 2020, a group from University of Saskatchewan along with researchers from
National Institute of Technology, Trichy, India and International Road Dynamics, Canada
published their research with COVID CXRs and a CheXNet based Deep CNN called CXNet [8].
The research team also published their base model’s performance, which was comparable to
CheXNet even though it had a significantly lesser number of parameters. To achieve this, the
team used various data pre-processing techniques like, lung segmentation [11] and CLAHE and
BEASF [25,16] on CXRs. They were able to improve on some fault points of CheXnet. CheXnet
was often found distracted by the textual data on images, a problem which the team tackled by
lung segmentation using U-nets, forcing the model to focus on the required ROI. Their final
model, Hierarchical Multi-class COVID-CXNet, achieved an F1 score of 0.94 for binary
classification and 0.85 for 3 class classification. Although their model performed very well, they
pointed out a major problem of lack of data for training. They trained their model on only about
7700 [9] images available for COVID detection.
Apart from the aforementioned techniques, researchers have also tried bone suppression [12, 13]
to increase the model performance. Bone suppression separates the soft tissues from the bones,
removing some distractions for the deep learning models. Researchers of this [12] paper proved
the efficacy of this process in detecting tuberculosis from CXRs.
Research Gaps
1. CheXNet was often found distracted by the textual data on X-ray images, which can be
solved by lung-segmentation.
2. University of Saskatchewan used Lung Segmentation, BEASF and CLAHE in their data
preprocessing pipeline, but they left out bone suppression.
3. Although the model presented by researchers at University of Saskatchewan performed
well for detecting Covid even with lesser number of parameters compared to other
researchers mentioned in their paper, they stated the lack of data as a limitation for the
bias in their model.
4. Another group of independent researchers tried out bone shadow suppression, but left out
histogram-equalization and lung segmentation in their data preprocessing pipeline, which
have proved to be effective in increasing model performance.
Objectives
1. Building upon the existing research on effectiveness of data preprocessing for medical
images, we want to find and implement an optimized data preprocessing pipeline that
achieves lung segmentation [11, 14] using neural networks like U-net, bone suppression
using an auto-encoders network and HE using techniques like BEASF, CLAHE, DHE
and others.
2. For increasing the model’s explainability, we want to implement the Grad-CAM [21]
algorithm, which would help determine if the model is looking at the correct places or
not.
3. For measuring model’s performance, we use F1 score for multi-class classification.
4. Lastly, we are proposing an Inception [19] based Deep CNN trained from scratch for this
application.
5. Along with our proposed model, we would compare the results with SqueezeNet based
and ResNet-34 deep learning models.
Research Methodology
1. Dataset preparation:
a. We are using the NIH Chest X-ray dataset [6] for our final disease classification.
b. We are combining the open-source COVID-19 dataset, “COVID-19 Image Data
Collection” [9] compiled by a team from IEEE.
c. For the Segmentation model, we are using datasets of Montgomery County chest
X-ray set (MC) and Shenzhen chest X-ray set from U.S. National Library of
Medicine [10].
d. For Bone suppression, we are using the BSE-JSRT dataset [13].
2. Bone Suppression Model Training: we’re using an auto-encoder based model for bone
suppression [22].
3. HE: for this step, we would use OpenCV [18] and python’s NumPy library.
4. Lung Segmentation: for this we’re using a U-Net model which has been proven to be
very effective in medical segmentation tasks [26], along with edge-dilation to preserve
the lung structure better.
5. In the final stage, we have chosen a deep learning model which would accept inputs from
the preprocessing pipeline and the reshaped x-ray image for context preserving [19]. For
the outputs, we would have a primary output path for pulmonary pathology classification
using softmax function and a secondary output for binary classification using a sigmoid
function for COVID-19 dataset. Primarily, we are using an inception architecture to keep
the number of parameters low while giving attention to different levels of details in the
x-ray images.
6. Along with the Inception based model, we would be using a SqueezeNet and a ResNet-34
model to compare the performance.
Methodology of Study
Objective-1:
1) Datasets :- (50% complete) We have collected the datasets required for the project,
details of whom are mentioned in the table below:
Atelectasis 5789
Cardiomegaly 1010
Effusion 6331
Infiltration 10317
Mass 6046
Nodule 1971
Pneumonia 1062
Pneumothorax 2793
Normal 84312
Total 119631
For Bone Suppression dataset: One hundred and fifty-four conventional chest radiographs with a
lung nodule and 93 radiographs without a nodule were selected from 14 medical centers and
were digitized by a laser digitizer with a 2048 x 2048 matrix size (0.175-mm pixels) and a 12-bit
gray scale. It was developed by the JSRT (Japanese Society of Radiological Telhnology).
Bone Suppression
2) Histogram Equalization: For this we tested three algorithms (CLAHE, BEASF and
DHE):
From the results in this image, we can see, CLAHE performs the best among the tested
algorithms.
Objective-4 (50% done)
1) Lung Segmentation: We have created all but one of the scripts for final testing. Below is
the model architecture.
ModuleList1: - -
- DoubleConv 2-1 - -
- Sequential 3-1 [-1, 64, 512, 512] 37,696
ModuleList1: - -
- DoubleConv 2-2 - -
- Sequential 3-2 [-1, 128, 256, 256] 221,696
ModuleList1: - -
- DoubleConv 2-3 - -
- Sequential 3-3 [-1, 256, 128, 128] 885,760
ModuleList1: - -
- DoubleConv 2-4 - -
- Sequential 3-4 [-1, 512, 64, 64] 3,540,992
DoubleConv: 1-5 - -
- Sequential 2-5 - -
- Conv2D 3-5 - 4,718,592
- BatchNorm2D 3-6 - 2,048
- ReLU 3-7 - -
- Conv2D 3-8 - 9,437,184
- BatchNorm2d 3-9 - 2,048
- ReLU 3-10 [-1, 1024, 32, 32] -
ModuleList: 1 - -
- ConvTranspose2d: 2-6 - 2,097,664
- DoubleConv: 2-7 - -
- Sequential: 3-11 [-1, 512, 64, 64] 7,079,936
- ConvTranspose2d: 2-8 - 524,544
- DoubleConv: 2-9 - -
- Sequential: 3-12 - 1,770,496
- ConvTranspose2d: 2-10 - 131,200
- DoubleConv: 2-11 - -
- Sequential: 3-13 [-1, 256, 128, 128] 442,880
- ConvTranspose2d: 2-12 - 32,832
- DoubleConv: 2-13 - -
- Sequential: 3-14 [-1, 64, 512, 512] 110,848
While testing the model with just 1 image pair, we got these results:
2) Bone Suppression: We have created all but one of the scripts for model testing. Below is the
model architecture:
a) Inception based model: The model shown here is not the final, as one of the fully
connected layers currently has 96% of the model’s parameter weight. We are working on
updating this architecture.
Limitations:
1) Mainly, our model’s performance would be limited by the lack of quality data for the
bone suppression module.
2) Another limitation is the computational power required to train the model efficiently.
3) We are currently fixing the input size of the modules to 512x512, with more
computational capacity we would be able to increase the image input size, hence
increasing the information the model has to work with.
Gantt chart
References
[1] “RSNA Pneumonia Detection Challenge,” Kaggle.com. [Online]. Available:
https://www.kaggle.com/c/rsna-pneumonia-detection-challenge. [Accessed:
04-Sep-2021].
[5] B. Kelly, “The chest radiograph,” Ulster Med. J., vol. 81, no. 3, pp. 143–148, 2012.
[10] S. Jaeger, S. Candemir, S. Antani, Y.-X. J. Wáng, P.-X. Lu, and G. Thoma, “Two public
chest X-ray datasets for computer-aided screening of pulmonary diseases,” Quant.
Imaging Med. Surg., vol. 4, no. 6, pp. 475–477, 2014.
[11] S. Rajaraman, L. Folio, J. Dimperio, P. Alderson, and S. Antani, “Improved semantic
segmentation of tuberculosis-consistent findings in chest X-rays using augmented
training of modality-specific U-Net models with weak localizations,” arXiv [cs.CV],
2021.
[12] S. Rajaraman, G. Zamzmi, L. Folio, P. Alderson, and S. Antani, “Chest X-ray bone
suppression for improving classification of tuberculosis-consistent findings,”
Diagnostics (Basel), vol. 11, no. 5, p. 840, 2021.
[14] Y. Gordienko et al., “Deep learning with lung segmentation and bone shadow exclusion
techniques for chest X-ray analysis of lung cancer,” arXiv [cs.LG], 2017.
[16] J. Ma, X. Fan, S. X. Yang, X. Zhang, and X. Zhu, “Contrast limited adaptive histogram
equalization based fusion for underwater image enhancement,” Preprints, 2017.
[19] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the inception
architecture for computer vision,” in 2016 IEEE Conference on Computer Vision and
Pattern Recognition (CVPR), 2016.
[22] D. Y. Oh and I. D. Yun, “Learning bone suppression from dual energy chest X-rays
using adversarial networks,” arXiv [cs.CV], 2018.
[24] “India: air pollution deaths by type 2019,” Statista.com. [Online]. Available:
https://www.statista.com/statistics/1194824/india-air-pollution-deaths-by-type/.
[Accessed: 04-Sep-2021].