Professional Documents
Culture Documents
MI-UNet_Multi-Inputs_UNet_Incorporating_Brain_Parcellation_for_Stroke_Lesion_Segmentation_From_T1-Weighted_Magnetic_Resonance_Images
MI-UNet_Multi-Inputs_UNet_Incorporating_Brain_Parcellation_for_Stroke_Lesion_Segmentation_From_T1-Weighted_Magnetic_Resonance_Images
2, FEBRUARY 2021
Authorized licensed use limited to: University of Sargodha. Downloaded on November 06,2023 at 09:46:53 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: MI-UNet: MULTI-INPUTS UNet INCORPORATING BRAIN PARCELLATION FOR STROKE LESION SEGMENTATION 527
1
CNN is most commonly applied to analyzing visual imagery N
because it can automatically extract image features. UNet has LCE = − (gn · log(pn ) + (1 − gn ) · log(1 − pn )). (6)
N n=1
an encoder-decoder style that solves segmentation tasks in an
end-to-end manner [36]. As illustrated in Fig. 4, the proposed 2) Training Scheme: To feed 3D data into a 2D MI-UNet,
MI-UNet can be conducted in both 2D and 3D settings. we use their 2D slices along the axial plane. As shown in Fig. 4,
For each subject, its brain parcellation L1 and T1w MR image the 2D brain parcellation and T1w MR image slices are fed into
I1 are sliced/sampled and jointly fed into 2D/3D MI-UNet as a MI-UNet jointly as a two-channel input with a size of 233 ×
two-channel input. The hidden layers of UNet/MI-UNet can be 197 × 2. The 2D MI-UNet is trained for 50 epochs.
reasonably categorized into two parts: (1) an encoder part that We train our proposed 3D MI-UNet for 30 epochs, with each
can capture multi-level, multi-resolution features by computing one again composed of 20 sub-epochs. At each sub-epoch, 100
feature maps at different image resolutions, and (2) a decoder samples with a size of 49 × 49 × 49 × 2 are randomly selected
part that can localize stroke lesions precisely by computing the from the training images and the corresponding ground truth
feature representations at different scales. The structure details segmentation images [20]. As shown in Fig. 4, we use 3D
Authorized licensed use limited to: University of Sargodha. Downloaded on November 06,2023 at 09:46:53 UTC from IEEE Xplore. Restrictions apply.
530 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, VOL. 25, NO. 2, FEBRUARY 2021
TABLE I
ARCHITECTURES OF THE PROPOSED 2D MI-UNET AND 3D MI-UNET
a
“3 × 3, 64 conv” corresponds to a Conv-BN-ReLU layer with a kernel size of 3 × 3 and 64 features. “[ ]×d” denotes that the convolution block is repeated for d times.
b
[ ]-[ ] denotes concatenation of two feature maps. C(convolution 4) denotes the central-cropping operation to guranttee the same size in the concatenation operation.
MR image patches and the corresponding 3D brain parcellation normalized to be 1 mm3 voxel resolution and all volumes are of
patches as the two-channel inputs to 3D MI-UNet. the same size 233 × 197 × 189. In terms of stroke type found
3) Inference Scheme: At the testing stage of 2D MI-UNet, in the ATLAS dataset, most cases are embolic. All experiments
we use the axial slices of brain parcellation and T1w MR image are conducted in a five-fold cross-validation manner. For data
as the input to the trained 2D model. For 3D MI-UNet, brain par- partitioning, we first index all 229 cases according to their names
cellation and T1w MR image go through the trained 3D U-Net and then separate the training data and testing data according to
via sliding-window sampling. Finally, we perform thresholding their indices. More specifically, the data having indices {1, 6,
processing on all prediction outcomes of 2D/3D MI-UNet. Con- 11, ..., 226}, {2, 7, 12, ..., 227}, {3, 8, 13, ..., 228}, {4, 9, 14, ...,
cretely speaking, if the probability that a pixel/voxel is predicted 229}, {5, 10, 15, ..., 225} are respectively used for testing for
to be foreground is larger than 0.5, its label will be set to be one, fold 1 to fold 5 and the others are used for training. As shown in
and otherwise zero. Fig. 5, the testing data in every partition cover different lesion
sizes. Through a five-fold cross-validation, every data is used
D. Implementation Details as the testing data once. We also perform Wilcoxon signed-rank
We have made the LDDMM-image algorithm publicly avail- tests to quantify the group differences between every two folds
able at GitHub.1 We also made the brain parcellation used in this in terms of lesion sizes. There is no statistically significant group
work publicly available at Github.2 The implementations of 2D difference (p-value > 0.05).
UNet and 3D UNet were based on the work of Li et al.3 [17],
Wu et al. [20] and Dolz et al. [39].4 The batch size was set to B. Evaluation Metrics
be 8 for 2D MI-UNet and 32 for 3D MI-UNet. Adam optimizer The Dice score is an evaluation metric used to quantify the
was used and the learning rate was set to be 0.001. overlapping ratio between two binary sets, which is defined as
2TP
III. DATASET AND EVALUATION METRICS Dice = , (7)
2TP + FP + FN
A. Dataset where TP, FP, and FN respectively represent the pixel counts of
We evaluate our approach on the ATLAS dataset [7], con- true positives (model correctly predicts), false positives (model
sisting of 229 subjects. All images have been resampled and mis-classifies to be positive), and false negatives (model mis-
classifies to be negative). In medical image segmentation tasks,
1 [Online].
there is usually a strong size imbalance between foreground and
Available: https://github.com/KwuJohn/lddmmMASGD
2 [Online]. Available: https://github.com/sustecher/StrokeSegmentation
background, and the background typically occupies over 95% of
3 [Online]. Available: https://github.com/hongweilibran/wmh_ibbmTum the whole volume [38]. But the foreground is what we care more,
4 [Online]. Available: https://github.com/josedolz/LiviaNET namely our object of interest. As such, here “correctly predicts”
Authorized licensed use limited to: University of Sargodha. Downloaded on November 06,2023 at 09:46:53 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: MI-UNet: MULTI-INPUTS UNet INCORPORATING BRAIN PARCELLATION FOR STROKE LESION SEGMENTATION 531
Fig. 5. Lesion size analysis of each testing data partition in the five-fold experiment.
Authorized licensed use limited to: University of Sargodha. Downloaded on November 06,2023 at 09:46:53 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: MI-UNet: MULTI-INPUTS UNet INCORPORATING BRAIN PARCELLATION FOR STROKE LESION SEGMENTATION 533
Fig. 9. Representative segmentation results obtained from 3D UNet and 3D MI-UNet. The first row shows the 3D displays of the ground truth
and all automated segmentations. From the second to the fourth row, MR image and the corresponding segmentation are displayed at three planes
(axial, coronal, and sagittal).
TABLE III
COMPARISONS OF DIFFERENT STROKE LESION SEGMENTATION RESULTS OBTAINED ON THE ATLAS DATASET. ↓ INDICATES THAT THE SMALLER THE BETTER
methods. We would like to kindly point out that all the results evaluation metrics, are statistically significant (p-value < 0.05).
used for comparison are directly obtained from their published From Table III, we find that generally deep learning-based
papers. In other words, we do not implement their methods at methods outperform classical segmentation techniques because
our end, as we believe a direct comparison with the best results they can automatically capture discriminative features that out-
reported by the authors themselves would be the fairest way. In perform hand-crafted features [47]. With respect to Dice, the
that table, the results of the four classical methods (Cluster, ALI, proposed method has the highest mean score of 49.45%/56.72%
lesion_gnb, and LINDA) were obtained from Ito et al. [6], the among 2D/3D deep learning-based methods. Meanwhile, our
results of the three 2D deep learning-based methods (ResUNet, best model (3D MI-UNet) has a smaller average HD than LINDA
2D Dense-UNet, X-Net) were reported by Qi et al. [15], and (from 42.07 mm to 23.94 mm) [11] and a smaller ASSD (from
the results of MSDF-Net and D-UNet were reported by Zhou 10.49 mm to 7.00 mm) than ALI [9]. The performance of the
et al. [25] and Liu et al. [46]. Cluster method is the best in terms of recall, but its Dice and
Our goal is to show that the additional brain parcellation can precision are the lowest, indicating that Clusterize misclassifies
improve the segmentation accuracy over baseline models (for a lot of non-lesion areas to be lesions, and hence yields the best
both 2D UNet and 3D UNet). As such, we conduct the 2D/3D recall. All deep learning-based methods have relatively lower
U-Net and MI-UNet and obtain results by ourselves (the last recall but higher precision and Dice than traditional methods. In
four rows in Table III). Most of the improvements (2D MI-UNet summary, the proposed strategy can significantly improve the
vs. 2D UNet, 3D MI-UNet vs. 3D UNet) in terms of the five segmentation performance of both 2D UNet and 3D UNet and
Authorized licensed use limited to: University of Sargodha. Downloaded on November 06,2023 at 09:46:53 UTC from IEEE Xplore. Restrictions apply.
534 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, VOL. 25, NO. 2, FEBRUARY 2021
Authorized licensed use limited to: University of Sargodha. Downloaded on November 06,2023 at 09:46:53 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: MI-UNet: MULTI-INPUTS UNet INCORPORATING BRAIN PARCELLATION FOR STROKE LESION SEGMENTATION 535
[13] Y. Zhang, J. Wu, C. Wanli, C. Yifan, and X. Tang, “Prostate segmentation [31] J. Wu and X. Tang, “A large deformation diffeomorphic framework for
using Z-Net,” in Proc. Int. Symp. Biomed. Imag., 2019, pp. 11–14. fast brain image registration via parallel computing and optimization,”
[14] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional networks Neuroinformatics, vol. 18, pp. 251–266, 2020.
for biomedical image segmentation,” in Proc. Med. Image Comput. Com- [32] X. Qian, Y. Lin, Y. Zhao, X. Yue, B. Lu, and J. Wang, “Objective ven-
put. Assist. Intervention, 2015, pp. 234–241. tricle segmentation in brain ct with ischemic stroke based on anatomical
[15] K. Qi et al., “X-Net: Brain stroke lesion segmentation based on depthwise knowledge,” BioMed Res. Int., vol. 2017, 2017, Art. no. 8690892.
separable convolution and long-range dependencies,” in Proc. Med. Image [33] M. I. Miller, A. Trouvé, and L. Younes, “On the metrics and euler-lagrange
Comput. Comput. Assist. Intervention, 2019, pp. 247–255. equations of computational anatomy,” Annu. Rev. Biomed. Eng., vol. 4,
[16] H. Yang et al., “CLCI-Net: Cross-level fusion and context inference no. 1, pp. 375–405, 2002.
networks for lesion segmentation of chronic stroke,” in Proc. Med. Image [34] B. B. Avants, C. L. Epstein, M. Grossman, and J. C. Gee, “Symmetric
Comput. Comput. Assist. Intervention, 2019, pp. 266–274. diffeomorphic image registration with cross-correlation: Evaluating auto-
[17] H. Li et al., “Fully convolutional network ensembles for white matter mated labeling of elderly and neurodegenerative brain,” Med. Image Anal.,
hyperintensities segmentation in MR images,” NeuroImage, vol. 183, vol. 12, no. 1, pp. 26–41, 2008.
pp. 650–665, 2018. [35] J. Wu, Y. Zhang, and X. Tang, “Simultaneous tissue classification and
[18] G. Zeng and G. Zheng, “3D tiled convolution for effective segmentation lateral ventricle segmentation via a 2D U-Net driven by a 3D fully
of volumetric medical images,” in Proc. Med. Image Comput. Comput. convolutional neural network,” in Proc. Eng. Med. Biol. Soc., 2019,
Assist. Intervention, 2019, pp. 146–154. pp. 5928–5931.
[19] K. C. Wong and M. Moradi, “SegNAS3D: Network architecture search [36] T. Falk et al., “U-Net: Deep learning for cell counting, detection, and
with derivative-free global optimization for 3D image segmentation,” morphometry,” Nature Methods, vol. 16, no. 1, pp. 67–70, 2019.
in Proc. Med. Image Comput. Comput. Assist. Intervention, 2019, [37] F. Isensee et al., “nnU-Net: Self-adapting framework for U-Net-based
pp. 393–401. medical image segmentation,” 2018, arXiv:1809.10486.
[20] J. Wu, Y. Zhang, and X. Tang, “A multi-atlas guided 3D fully convolutional [38] F. Milletari, N. Navab, and S.-A. Ahmadi, “V-Net: Fully convolutional
network for MRI-based subcortical segmentation,” in Proc. Int. Symp. neural networks for volumetric medical image segmentation,” in Proc.
Biomed. Imag., 2019, pp. 705–708. Int. Conf. 3D Vis., 2016, pp. 565–571.
[21] Ö. Çiçek, A. Abdulkadir, S. S. Lienkamp, T. Brox, and O. Ronneberger, [39] J. Dolz, C. Desrosiers, and I. B. Ayed, “3D fully convolutional networks
“3D U-Net: Learning dense volumetric segmentation from sparse an- for subcortical segmentation in MRI: A large-scale study,” NeuroImage,
notation,” in Med. Image Comput. Comput. Assist. Intervention, 2016, vol. 170, pp. 456–470, 2018.
pp. 424–432. [40] L.-C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, “Encoder-
[22] K. Kamnitsas et al., “Efficient multi-scale 3D CNN with fully connected decoder with atrous separable convolution for semantic image segmenta-
CRF for accurate brain lesion segmentation,” Med. Image Anal., vol. 36, tion,” in Proc. Eur. Conf. Comput. Vis., 2018, pp. 801–818.
pp. 61–78, 2017. [41] W. Chen et al., “Prostate segmentation using 2D bridged U-Net,” in Proc.
[23] O. Maier et al., “ISLES 2015—A public evaluation benchmark for is- Int. Joint Conf. Neural Netw., 2019, pp. 1–7.
chemic stroke lesion segmentation from multispectral MRI,” Med. Image [42] M. Drozdzal, E. Vorontsov, G. Chartrand, S. Kadoury, and C. Pal, “The
Anal., vol. 35, pp. 250–269, 2017. importance of skip connections in biomedical image segmentation,” in
[24] R. Zhang et al., “Automatic segmentation of acute ischemic stroke from Deep Learning and Data Labeling for Medical Applications. Berlin,
DWI using 3D fully convolutional DenseNets,” IEEE Trans. Med. Imag., Germany: Springer, 2016, pp. 179–187.
vol. 37, no. 9, pp. 2149–2160, Sep. 2018. [43] L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille,
[25] Y. Zhou, W. Huang, P. Dong, Y. Xia, and S. Wang, “D-UNet: A “Deeplab: Semantic image segmentation with deep convolutional nets,
dimension-fusion U shape network for chronic stroke lesion segmenta- atrous convolution, and fully connected CRFs,” IEEE Trans. Pattern Anal.
tion,” IEEE/ACM Trans. Comput. Biol. Bioinformat., to be published. Mach. Intell., vol. 40, no. 4, pp. 834–848, Apr. 2018.
[26] S. Winzeck et al., “ISLES 2016 and 2017-benchmarking ischemic stroke [44] Z. Zhang, Q. Liu, and Y. Wang, “Road extraction by deep residual U-
lesion outcome prediction based on multispectral MRI,” Frontiers Neurol., Net,” IEEE Geosci. Remote. Sens. Lett., vol. 15, no. 5, pp. 749–753, May
vol. 9, 2018. 2018.
[27] L.-P. Berner et al., “MRI assessment of ischemic lesion evolution within [45] X. Li, H. Chen, X. Qi, Q. Dou, C.-W. Fu, and P.-A. Heng, “H-DenseUNet:
white and gray matter,” Cerebrovascular Diseases, vol. 41, no. 5-6, Hybrid densely connected UNet for liver and tumor segmentation from
pp. 291–297, 2016. CT volumes,” IEEE Trans. Med. Imag., vol. 37, no. 12, pp. 2663–2674,
[28] C. Matute, M. Domercq, A. Pérez-Samartín, and B. R. Ransom, “Protect- Dec. 2018.
ing white matter from stroke injury,” Stroke, vol. 44, no. 4, pp. 1204–1211, [46] X. Liu et al., “MSDF-Net: Multi-scale deep fusion network for stroke
2013. lesion segmentation,” IEEE Access, vol. 7, pp. 178 486–178 495, 2019.
[29] Y. Wang, G. Liu, D. Hong, F. Chen, X. Ji, and G. Cao, “White matter injury [47] I. Razzak, M. Imran, and G. Xu, “Efficient brain tumor segmentation
in ischemic stroke,” Prog. Neurobiol., vol. 141, pp. 45–60, 2016. with multiscale two-pathway-group conventional neural networks,” IEEE
[30] C. Chen, A. Bivard, L. Lin, C. R. Levi, N. J. Spratt, and M. W. Parsons, J. Biomed. Health Inform., vol. 23, no. 5, pp. 1911–1919, Sep. 2019.
“Thresholds for infarction vary between gray matter and white matter in [48] X. Yang, R. Kwitt, M. Styner, and M. Niethammer, “Quicksilver: Fast
acute ischemic stroke: A CT perfusion study,” J. Cerebral Blood Flow predictive image registration—A deep learning approach,” NeuroImage,
Metabolism, vol. 39, no. 3, pp. 536–546, 2019. vol. 158, pp. 378–396, 2017.
Authorized licensed use limited to: University of Sargodha. Downloaded on November 06,2023 at 09:46:53 UTC from IEEE Xplore. Restrictions apply.