Professional Documents
Culture Documents
Lecture 10 Merged
Lecture 10 Merged
• Image segmentation:
– Detection and delineation of objects of interest in the images
– Prerequisite for morphometry and regional quantification
• Image registration:
Normal
– Spatial alignment of different images (different modalities / time points / subjects…)
Stroke
fMRI – Prerequisite for fusion and joint analysis of complementary image information
6
• Image visualization:
– Presentation of the relevant information that was extracted from the images
– Prerequisite for clinical interpretation
0
Huntington
3 4
• Complex data • Incorporate prior knowledge about the appearance of the objects of interest in the images
– multi-dimensional, multi-temporal, multi-modal, multi-subject – photometric properties (intensity, contrast, texture…)
– limited image quality: resolution, contrast, noise, artifacts è ambiguity – geometric properties (position, shape, deformations...)
• Complex objects – context (other objects, clinical information…)
– 3D anatomical shapes
– normal biological variability • Model = mathematical representation of prior knowledge
– abnormalities (pathology…) – must be flexible (= parameterized) to account for variability
• Complex applications – can itself be represented as an image (e.g. an atlas)
– continuous technological advances in medical imaging
– increasing clinical requirements • Image analysis problem formulated as an optimization problem
• Complex validation – Find the model parameters that “best fit” the image data
– lack of objective ground truth in clinical images – Suitable criterion to measure “goodness of fit” (= objective function)
– observer variability – Suitable computational strategy to find the optimal solution (= optimization method)
5 6
Maximum A Posteriori Probability (MAP) formulation Approach 1: energy minimization
Prob(M(Q)) = prior probability of the model with parameters Q Energy minimization problem
Prob(I | M(Q)) = data likelihood for the specified model parameters è Flexibility to define the energy terms…
Prob(I ) = prior probability of observing image I (independent of Q ) è Based on heuristics, physics, statistics…
Prob(M(Q) | I) = posterior probability of the model with parameters Q for the observed image I
g = user-specified weight (hyper-parameter)
è Tunable behavior…
7 8
M = contour with N landmarks Eint (Q): shape model M = atlas Eint (Q): regularization
Q = landmark coordinates Eext (I|Q): intensity model 9 Q = deformation field Eext (I|Q): similarity measure 10
MAP formulation
F
e O
Estimate Prob(M(q)|I ) directly from training data a u
(supervised classification)
t t
Data Classifier
F = classifier / regressor u p
Maps given input image I onto most likely model r u
instance q* based on previous training data
e t
s
Instead of full data I: feature vector f
(= dimensionality reduction) SVM
Domain Handcrafted Random forest
knowledge feature Neural network
11 vector 12
Deep learning: automated optimal features A single neural node
x0 = 1 w0
O
u a = f(z)
w1
Classifie t x1 z = Si wi*xi f
Data Features
r p
u Linear Non-linear
t weighted sum activation function
w2
a
1
Deep neural network x2 sigmoid:
Domain
with automated discovery of
knowledge 0
optimal features z
13 14
z = 0 è a = 0.5
Rationale Rationale
x2 x2
z<0èaà0 z = 0 è a = 0.5
z = Si wi*xi z = Si wi*xi
a = Prob(Class = 1)
z>0èaà1
z>0èaà1 x1 z<0èaà0 x1
15 16
1 if Normal,
0 otherwise
1 if Abnormal type A,
0 otherwise
1 if Abnormal type B,
0 otherwise
r-1 i1 i2 i3 w1 w2 w3 r-1
r i4 i0 i5 * w4 w0 w5 r j
r+1 i6 i7 i8 w6 w7 w8 r+1
Shorthand:
sigmoid:
1 tanh: 1
0 z -1
MOST POPULAR a
a
Rectified Linear Unit: Leaky RELU:
(RELU) (LRELU)
0 z z
0
27 28
29 30
Training of deep CNNs: plenty of heuristics Drop out
31 32
33 34
35 36
Example: AlexNet for object recognition (2012) CNN variants
37 38
Christies, 2018
StarGAN, 2018
The DL zoo RN
N
LSTM
t GA
U-Ne N
SqueezeNet Deep learning for medical image computing
U-net, Ronneberger et al.
Part 2: Applications
dic
p Me
Dee
41 42
General principles of deep learning Deep Learning for Computer Vision
43 44
Deep Learning for Medical Imaging Deep Learning for Image Segmentation
Reconstruction & QC Quantification Treatment planning
Diagnosis © ADNI
Outcome prediction
Ubiquitous in medical image computing
CNNs for image segmentation: U-Net CNNs for image segmentation: DeepMedic
V(1&2)
Loss function
V (1& 2)
DSC =
(cross-entropy, Dice) (V (1) +V (2)) / 2
Image transformations
x mm over-segmentation: DSC = 82%
Class weighted sample generator
49 x mm over-segmentation: DSC = 90% 50
Sample transformations
Example 1: Delineation of left ventricle in cardiac MRI Example 2: Delineation of left ventricle in cardiac US
• Unet-based architecture • Unet-based architecture LV
– Input = stack of 11 T1-weighted images – Input = 2CH US, ED or ES
(mid-cavity SA, 320 x 320, pre-aligned) (672 x 672, point of cone aligned)
– Output = segmentation of myocardium – Output = delineation of LV epi/endo + LA
• Training: 146 patients x 5 slices • Training: 450 patients x 2 phases
– Loss-function = Dice coefficient – Loss-function = mean DSC
Myocard
– No data augmentation – No data augmentation
– Trained from scratch (~ 30’) – Trained from scratch
• Test: 50 patients x 5 slices • Test: 50 patients x 2 phases
• From concept to state-of-the-art result: < 2 weeks…
Mean DSC [%] 91.8
MAD_Endo [mm] 0.94 LA
DSC (%) Mean +/- Std
MAD_Epi [mm] 1.36
LV 91.1 +/- 4.9
HD_Endo [mm] 2.54
Expert Myocard 82.2 +/- 7.7
AI tool HD_epi [mm] 3.40
Best DSC = .97 Median DSC = .93 Worst DSC = .40 DSC = .80
LA 84.7 +/- 15.4 Max Median Min
51 52
Example 3: Brain tissue segmentation (2D CNN) Example 4: Brain tissue segmentation (3D CNN)
Moeskops et al., Wachinger et al.,
IEEE Trans Medical Imaging, 2016 DeepNAT, NeuroImage 2018
25 structures
Prenatal, young adults, aging subjects T1-weighted, 1 mm3
Modalities: T1, T2 13x13x13 patches
Very limited training set (5-10 images) 2.7 million parameters
2D patches at multiple scales
Use coordinates for local context
Preprocessing: BFC, brain mask
No post-processing 30 images (20 training, 10 test)
Performance: DSC 80-90%
ChildBrain 2019 53 54
Example 5: Detection and segmentation of brain pathology FLAIR T1 T1 CE T2 GT Prediction
Tumor
MS
WHOLE TUMOR:
© BRATS, ISLES
GT Prediction
55
GT Prediction
GT Prediction GT Prediction
FLAIR T1 T1 CE T2 GT Prediction FLAIR T1 T1 CE T2 GT Prediction
GT Prediction GT Prediction
Conventional approach:
perfusion model
15 s 20 s 25 s
Predicted
Multiclass classification Tmax
ENHANCING TUMOR:
Alternative approach: Ground truth
data-driven prediction based on MRI
GT Prediction
64
Left Cochlea
Planning CT Parotid glands
Brainstem Right Cochlea
Upper Esophagus
CT acquisition Contouring RT Plan optimization Final dose calculation
Glottic Larynx
Mandible
Oral Cavity
Supraglottic Larynx
Mandible Left Parotid gland
Right Parotid gland
Inferior PCM
Mid PCM
Superior PCM
Left Submandibular gland
Manual or semi-automatic Manual interventions One patient specific treatment Oral cavity Spinal cord
plan Right Submandibular gland
Target volumes Organs at risk
Spinal Cord
45min – 2hours hours
H&N : 16 OAR, clinical guidelines (Brouwer et al, 2016)
Inter- and intra- Inter-institutional Semi-automated delineation approaches have been developed (e.g. atlas-based), but complicated to use
observer variations
65 In clinical practice: manual delineation…
variability 66
DeepVoxNet 3D CNN for H&N OAR segmentation Initial results
DeepVoxNet
16 outputs
Loss function: = probability map
mean Dice per structure
è post-processing
~70 training images: planning CT + delineations exported from clinical TPS Expert
Trained from scratch (randomly initialized weights) AI tool
Adam optimizer with drop out
67 68
Integration in the clinical RT workflow in UZ Leuven Retrospective validation: AI tool mimicking the expert
Radiation Oncology Medical Imaging Research Center DSC (%) 10 test cases
Prospective validation: expert correcting the AI tool Clinical benefits of auto-delineation using DL ?
DSC (%) Retrospective Prospective
(10 cases) (20 cases) Experimental automated workflow
Brainstem 84.7 91.5
Left Cochlea 70.1 75.4
Right cochlea 80.1 73.1 Expert 1
Corrected
Upper esophagus 66.7 34.8
delineations
Glottic Larynx 53.3 39.4
Automated Expert 2
Mandible 84.9 95.9
delineations
Oral Cavity 67.5 83.5
(16 OAR)
Supraglottic Larynx 54.3 71.2
Left Parotid gland 80.8 86.3 Conventional manual workflow
Right Parotid gland 79.8 89.7
AI tool (5’) Corrections (15’) Inferior PCM 48.0 57.9
15 consecutive Clinical planning CT Expert 1
Manual
Mid PCM 55.5 60.9
H&N RT patients delineations
Prospective = ‘good enough!’ Superior PCM 36.5 46.1 Expert 2
Observer focuses on clear errors, Left Submandibular gland 69.6 78.8
irrelevant variability ignored
Right Submandibular gland 71.9 87.7
Direct clinical feedback:
Spinal Cord 81.9 95.9 71 72
mouth open/closed, missing organs
Auto-delineation performance: efficiency Auto-delineation performance: observer variability
Expert 1
Expert 2
Delineation time
= time spent by the expert to delineate or to correct Automated
all 16 OAR structures in one patient
Reduction of 38 %
73 74
Expert 1 Expert 1
Expert 2 Expert 2
Automated Automated
75 76
Beyond delineation in Radiology: computer-aided diagnosis Beyond delineation in RT: treatment planning
Cardiac MRI
Deep Learning today
Delineation Imaging biomarkers CT acquisition Contouring RT Plan optimization Final dose calculation
End-diastole End-systole
Epicardium Endocardium Epicardium Endocardium
Statistical modeling
Conventional
machine learning Feature selection
DL 1 DL 2 DL TPS
Plan
Images Contours Dose Images Actual dose
DVH constraints
Treatment setup
Sequential è sub-optimal use of correlation between CT / contours / dose Simultaneous, largely based on same image-derived features è exploit correlations
81 82
Pre-treatment planning CT On-board CBCT of today • Deep learning is based on deep convolutional neural networks
• A deep neural network is a NN with many hidden layers
Option 2:
• A deep NN aims at achieving a complex mapping of a high-dimensional input (an image) onto a user-specified output
Replan (classification, regression)
• Training involves optimizing the (very many) internal parameters of the NN (using back-propagation) by presenting samples
of input/output pairs until convergence (assessed by a suitable loss function)
• CNNs are conceptually simple, but due to the huge amount of parameters, many heuristics are needed to get them to work
Option 1: • CNNs are computationally complex and best implemented on GPUs
Adapt • CNNs were invented in the 1990s, but broke through 20 years later in computer vision due to cheaper GPUs, large image
databases on the internet and more clever training schemes
• DL using CNNs is a revolution in computer vision, as it works with rich image patterns directly instead of poor feature-based
representations derived from them
• Medical imaging applications are more complex, as data is scarce and specific expertise is needed
• Properly structured image annotations are key for deep learning
• The current AI hype is all about DL. The intelligence is in the combination of {data, annotations, domain knowledge}, not in
the algorithm itself.
87 88