Professional Documents
Culture Documents
Deep Learning For Fast Spatially Varying Deconvolution: Supplement
Deep Learning For Fast Spatially Varying Deconvolution: Supplement
This supplement published with Optica Publishing Group on 12 January 2022 by The Authors
under the terms of the Creative Commons Attribution 4.0 License in the format provided by the
authors and unedited. Further distribution of this work must maintain attribution to the author(s)
and the published article’s title, journal citation, and DOI.
Initialized filter
z = 1 µm
75 µm 75 µm
0 0
1 1
Learned filter
z = 300 µm
75 µm 75 µm
0 0
Fig. 1. Spatially varying PSFs. Measured PSFs at different Fig. 2. Learned and initialized filters. Central learned and
lateral positions for two different depth planes. Due to field- initialized filters at different depths used for the 3D reconstruc-
varying aberrations in the system, the PSFs change structure tion. For visualization, we scale the learned filters from 0 to 1.
across the field-of-view (FoV). Note that the images are con- In general, the filters can contain negative values. Note that
trast stretched to show detail. the images are contrast stretched.
D. Training and Testing Details within a certain threshold, the loss is below a certain threshold,
We train for 25 epochs for the 3D reconstruction case and 100 or the number of iterations exceeds a certain threshold.
epochs for the 2D case, using a learning rate of 1e−4 . We use
the ADAM optimizer [4] with default parameters throughout B. Learned filters
training. We allow the Wiener deconvolution filters and regu- We allow the network to learn the Wiener deconvolution filters.
larization parameters to update throughout training. We tested We initialize the network with a grid of measured filters, 3 × 3
having the Wiener filters and/or the regularization parameters for 2D and 3 × 3 × 32 for 3D, each is centered on a different
be fixed throughout training, but the best results were obtained region in the FoV. Fig. 1 shows initial filters at different positions
when they were allowed to change. Our loss function is an av- demonstrating the shift-variance of the system. By allowing the
erage of a structural similarity index measure (SSIM) loss and network to update the filters, the network can find filters that
L1 loss. For SSIM, we use an 11 × 11 Gaussian filter of width better approximate the shift-variance or better recover certain
1.5. We test the performance on both simulated and experimen- frequencies. Fig. 2, shows the initialized and learned filters. The
tal data. The experimental sample in main-text Figure 3(a) is a learned filters maintain similar structure in the central region
1951 USAF fluorescent resolution target, and in 3(b) is a freely as the initialized ones, while adding more information towards
moving stained tardigrade (water bear) captured at 40 frames the edges of the filter. At first glance, it is unclear if the extra
per second. information added to the learned filters is useful. However, by
examining the intermediate result of the Wiener deconvolution
step using both the initialized and learned filter in Fig. 3, we find
2. ADDITIONAL RESULTS that the deconvolved image using the learned filter has much
A. Timing results sharper features than the one using the initialized filter. This
allows the multiple learnable Wiener deconvolutions to provide
We compare our reconstruction time against FISTA with a
the CNN with sharp intermediate images that the CNN can
spatially-varying model and existing deep learning based mod-
further refine to remove low frequency artifacts. Note that we
els (U-Net, U-Net w/Wiener) by averaging over 100 runs on
do not require the learned filters to be positive.
GPU. The results are summarized in Table 1. We can see that
our method is significantly faster than existing spatially-varying
C. Deconvolution with localized PSFs
methods, and has comparable speeds to existing deep learning
based methods. For 2D, we can perform reconstructions at 30fps To demonstrate that our MultiWiener layer should generalize
for 336×480 images. For 3D, we can perform reconstructions at well to other imaging systems, we use Zemax to simulate PSFs
2.4fps for 336×480×32 volumes. This is significantly faster than from a traditional one-to-one imaging system with off axis aber-
what is possible without using deep learning-based methods rations (e.g. coma). Specifically, we simulate a dense grid of
(0.05 fps for 2D, 0.0015 fps for 3D), and could enable interactive 512 × 512 PSFs using the specifications of the 2D Miniscope
previewing. Timing results were performed on an RTX 2080 system [5]. From these PSFs, we simulate 2D measurements
Ti GPU. The stopping criterion for FISTA was chosen to avoid of sparse beads and a dense scene, of size 512 × 512, using the
unnecessary iterations. FISTA terminates if the solution is stable superposition principle. Fig. 4 shows the simulated measure-
ment and the results of performing Wiener deconvolution on
the measurement using the on-axis PSF and an off-axis PSF. As
expected, deconvolution with the on-axis PSF in a system with
Table 1. Reconstruction timing comparisons (GPU) spatially varying aberrations results in an image with good re-
construction quality only near the center portion of the image. In
FISTA U-Net WienerNet MultiWeinerNet
contrast, deconvolving with a PSF from an off-axis point results
2D 20s 0.021s 0.029s 0.032s in sharper image quality in the neighborhood from which this
PSF is sampled. This demonstrates how our MultiWiener stage
3D 665s 0.33s 0.34s 0.41s
can produce intermediate images where different portions of
Letter Optica 3
(a) Wiener deconvolution with initialized filter 3. Centre for Biomedical Image Analysis, “CytoPacq,” (https://cbia.fi.muni.
measurement initialized filter deconvolved image 1
cz/simulator).
4. D. P. Kingma and J. Ba, arXiv preprint arXiv:1412.6980 (2014).
5. K. K. Ghosh, L. D. Burns, E. D. Cocker, A. Nimmerjahn, Y. Ziv,
A. El Gamal, and M. J. Schnitzer, Nat. methods 8, 871 (2011).
75 µm
0
(b) Wiener deconvolution with learned filter
measurement learned filter deconvolved image 1
75 µm
0
the image are sharp depending on where the PSF was sampled.
These intermediate images can then be fed into the U-net, which
will combine and refine the images to produce a sharp image
across the FoV. We expect this general approach to work well in
a variety of imaging systems with spatially-varying aberrations.
REFERENCES
1. K. Yanny, N. Antipa, W. Liberti, S. Dehaeck, K. Monakhova, F. L. Liu,
K. Shen, R. Ng, and L. Waller, Light. Sci. & Appl. 9, 1 (2020).
2. G. Kuo, K. Monakhova, K. Yanny, R. Ng, and L. Waller, “Spatially-
varying microscope calibration from unstructured sparse inputs,” in
Computational Optical Sensing and Imaging, (Optical Society of Amer-
ica, 2020), pp. CF4C–4.
Letter Optica 4
FULL REFERENCES
1. K. Yanny, N. Antipa, W. Liberti, S. Dehaeck, K. Monakhova, F. L. Liu,
K. Shen, R. Ng, and L. Waller, “Miniscope3d: optimized single-shot
miniature 3d fluorescence microscopy,” Light. Sci. & Appl. 9, 1–13
(2020).
2. G. Kuo, K. Monakhova, K. Yanny, R. Ng, and L. Waller, “Spatially-
varying microscope calibration from unstructured sparse inputs,” in
Computational Optical Sensing and Imaging, (Optical Society of Amer-
ica, 2020), pp. CF4C–4.
3. Centre for Biomedical Image Analysis, “CytoPacq,” (https://cbia.fi.muni.
cz/simulator).
4. D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,”
arXiv preprint arXiv:1412.6980 (2014).
5. K. K. Ghosh, L. D. Burns, E. D. Cocker, A. Nimmerjahn, Y. Ziv,
A. El Gamal, and M. J. Schnitzer, “Miniaturized integration of a flu-
orescence microscope,” Nat. methods 8, 871–878 (2011).