Abstract
We present a data-driven approach to compensate for optical aberrations in calibration-free quantitative phase imaging (QPI). Unlike existing methods that require additional measurements or a background region to correct aberrations, we exploit deep learning techniques to model the physics of aberration in an imaging system. We demonstrate the generation of a single-shot aberration-corrected field image by using a U-net-based deep neural network that learns a translation between an optical field with aberrations and an aberration-corrected field. The high fidelity and stability of our method is demonstrated on 2D and 3D QPI measurements of various confluent eukaryotic cells and microbeads, benchmarking against the conventional method using background subtractions.
© 2020 Optical Society of America under the terms of the OSA Open Access Publishing Agreement
1. Introduction
Quantitative phase imaging (QPI) has rapidly emerged as a prominent imaging modality in life sciences and medicine owing to its non-invasive, label-free, and quantitative visualization at the individual single-cell level [1]. In general, QPI systems are implemented using a laser interferometry to measure optical phase delay induced by a transparent specimen, although several non-interferometric QPI techniques have also been investigated [2–5]. Due to its label-free and quantitative imaging capability, QPI has been widely used for diverse biomedical studies, including cell biology [6,7], infectious disease [8,9], microbiology [10,11], and therapeutics [12].
One of the important issues that can affect and deteriorate QPI imaging quality is aberration and coherent noise, resulted from the use of a coherent illumination source. Ideally, plane wave illumination that impinges on a sample is required for the precise measurement of the sample-induced optical phase delay [Fig. 1(a)]. However, in practice, the presence of optical aberration induces a deviation in the measured phase image from the correct one [Fig. 1(b)].
To address this issue, several approaches have been reported. First, employing spatially or temporally incoherent illumination could alleviate the coherent noise. However, the resulting short coherence length can degrade the performance of QPI and complicate its optical system [13–18]. Other straightforward methods to circumvent optical aberration in QPI are background subtraction [19] or sample translation [20]. In the background subtraction method [Fig. 1(c)], optical aberration is removed by subtracting the measured phase image of a sample from that of a background area without objects [Fig. 1(c)]. This background subtraction method is straightforward, but it requires additional measurement of an object-free background and is not always possible, particularly for confluent samples, including biological cell aggregates and tissue slices.
The sample translation method retrieves and removes optical aberration from two differential images with slight lateral shifts [20]. Although the sample translation method is not restricted to the confluency of samples, it still has a strict restriction that samples must not change during the shifts, which prevents probing the fast dynamics of biological phenomena. Alternatively, fitting with Zernike polynomials was reported [21]. Even though it does not require additional measurement, this fitting method only works for low-order aberration because of its assumption.
Here, we present a deep-learning-based aberration compensation for calibration-free QPI. Using a paired dataset (optical field with aberration and aberration-corrected field using the background subtraction method), we trained a deep convolutional neural network (DCNN) that learns the aberration patterns of an optical system and suppresses the aberration of input optical field images [Fig. 1(d)]. The proposed network, inspired by U-net [22], used an encoder-decoder network with skip connection in order to circumvent the bottleneck for information (Fig. 1(e); Appendix A and B) [23]. Our method enables the single-shot aberration-corrected measurement of an optical field. The main advantage of our approach is that confluent samples can be obtained via QPI because it does not require any additional images or assumptions. To demonstrate the applicability, we performed 2D and 3D QPI of various samples, including polystyrene microspheres and eukaryotic cells. We also verify the high-fidelity performance of the present method in comparison to the background subtraction method.
2. Results
2.1 Calibration-free 2D quantitative phase imaging
For experimental verification, we tested our method with unseen eukaryotic cells (HeLa, NIH-3T3, HEK-293T, MDA-231, and COS7 cells, See also Appendix C). The representative results of NIH-3T3 and Hek-293T are shown in Fig. 2. The input, output, and ground truth (GT) complex field images of an NIH-3T3 cell, and confluent HEK-293T cells are presented in Figs. 2(a) and 2(b), respectively. The background subtraction method was used as GT in this work. The aberrations in the optical input fields are effectively suppressed in the network outputs, and the results are compatible with those obtained using the conventional background subtraction method (GT in Fig. 2).
Parasitic fringe patterns, caused by multiple reflections, are suppressed. Intriguingly, the phase discontinuity caused by phase wrapping in the input phase (at bottom-right corner) is also clearly removed in the network output [Fig. 2(a)]. This implies that through the training of a paired data set (input and GT images), the algorithm automatically learns the information about an ideal imaging transfer condition, suppressing any deviation from an ideal optical imaging condition.
The capability of our method was further validated with HEK-293T cells that densely occupy the entire field of view (FoV) [Fig. 2(b) and Appendix D]. Similar to the NIH-3T3 case, the fringe noises and phase ambiguities were removed successfully, generating sufficiently clean images for probing biological information. This result indicates that our method can be effectively applied to confluent samples where a background region for calibration is very rarely found. On the other hand, it is difficult or even impossible to use the aberration correction methods that must exploit a background. We compared the difference between the GT and the network output image [Fig. 2(c)]. The cross-correlation and root mean square error (RMSE) were measured as 0.982 and 0.127 rad [NIH-3T3 cell, Fig. 2(a)], and 0.950 and 0.317 rad [HEK-293T cells, Fig. 2(b)], respectively. While negligibly blurred, the subcellular features such as nucleolus and vesicles are preserved well in the network output.
The network error level is compared with the coherent noise level of the optical system used to validate the proposed method quantitatively. The absolute phase error map E between the present method and the conventional background subtraction method was calculated as E = ϕGT – ϕOutput, where ϕGT and ϕOutput are the phase images of GT and the network output, respectively. The phase error map for the NIH-3T3 cell is shown in Fig. 2(c). The background region was obtained from the measured FoV by manually masking out the cell to compare the phase error with the coherent noise level [Fig. 2(d)].
In Fig. 2(e), the fractional histogram of the phase error E (blue) and coherent noise Ncoh in the background region (red) are presented with their standard deviations. Both centered at zero, the standard deviations (0.13 rad and 0.12 rad) are comparable, indicating that the performance of our approach is close to the fundamental upper limit. Similar quantifications for the entire dataset can be found in Appendix E, the mean cross-correlation is 0.93, the mean RMSE is 0.23 rad.
2.2 Calibration-free 3D quantitative phase imaging
The proposed method can be further utilized in 3D imaging. A three-dimensional refractive index (RI) tomogram is reconstructed from multiple 2D optical fields at various angle illuminations using the principle of inverse light scattering, which is known as optical diffraction tomography (ODT) [Fig. 3(a)]. ODT is a 3D QPI technique, and it reconstructs the 3D RI tomogram of a sample from multiple 2D optical field images by inversely solving Helmholtz equation [24].
In ODT, it is essential to correct the aberration of each optical field because varying aberration in the angle-scanned optical fields, which frequently occur in practice, can deteriorate the image quality of reconstructed tomograms [Fig. 3(b), green arrow].
Figures 3(c)–3(d) show how our method can significantly improve the performance of ODT, benchmarking against the background subtraction method. The aberration correction on 49 optical fields via our method leads to the substantially improved reconstruction of the tomogram, compared to the reconstructed tomogram from the optical fields with aberration. The subcellular structures of NIH-3T3 were clearly restored in our network output. The solid black arrows indicate that the morphology of organelles, including nucleoli (1), nuclear membrane (2), and lipid droplets (3), were successfully reconstructed, as done in the reconstructed tomogram using the background subtraction method (GT).
The results show that this method improves the applicability of QPI in two ways. First, the proposed method does not require any additional measurements, enabling ‘single-shot’ QPI. This aspect offers a robust phase measurement without sample restrictions such as staticity or sparsity. Second, our network can perform not only aberration correction but also a robust phase unwrapping process, which is difficult to conduct owing to its ill-posedness. The presented results reconfirm the use of deep learning for phase unwrapping, which has previously been reported in Ref. [25,26].
3. Conclusion
In sum, we have proposed and experimentally demonstrated a deep-learning approach that performs aberration correction. The deep-learning model learned mapping from the ‘input’ optical field with aberrations to ‘output’ aberration-corrected optical field using the background subtraction method. The model, trained with the paired dataset, generates aberration-corrected optical fields without any additional measurements, enabling single-shot QPI. We have verified the robust performance of our method by successfully operating 2D QPI on various eukaryotic cells and optical diffraction tomography where the 3D refractive index of a biological sample is reconstructed from the 2D aberration-corrected optical fields with high temporal stability (Appendix F). Recently, deep learning has been applied to segment background region in the measured FoV [27]. However, it necessitates object-free background areas in the image to model aberrations based on the Zernike functions, which is fundamentally different from our method. We believe our approach, capable of imaging confluent samples rapidly, can be used to investigate perplexing biomedical phenomena in diverse disciplines.
Several subsequent studies can be conducted to enhance the proposed method. First, advanced network architectures, such as Dual-CNN or LS-DNN [28,29], may mitigate some blurring and pixelization artifacts that have appeared in our results. Next, a more diverse dataset can further generalize our method. As we used a single QPI instrument and five types of eukaryotic cells in our work, our framework may suffer from generalization when dealing with unseen data distribution. Although the limited variability of aberration patterns is addressed with proper data augmentation techniques, it might be challenging to employ our trained network in the same QPI system with different aberration patterns or completely different QPI systems (Appendix G; Fig. 9). A generalization test across various devices will be helpful to improve the algorithm. Third, the data preparation process can also be simplified and automated by manufacturing permanent reference samples with various RI distributions [30]. Fourth, the present approach can be sequentially used or combined with a deep-learning-based phase retrieval algorithm [31–33], which can potentially even expedite the whole imaging and reconstruction process. Finally, the proposed method could be effectively integrated with various important applications of machine learning on QPI, such as classification and segmentation, and image reconstruction [34–36].
Appendix A: Network architecture for aberration compensation
The encoder extracts the features of the input image at various scales through the successive down-sampling convolutional blocks. The input image is first convolved with 4 × 4 filters with stride 2 and pad 1 to create 64 feature maps. Down-sampling blocks are successively applied, extracting feature maps of different levels. The blocks consist of leaky rectified linear units (Leaky Relu) [37], 4 × 4 convolutions with stride 2 and pad 1, and batch normalization [38], typically used to reduce internal covariate shift and training time. The number of convolutional filters doubles with each layer until it becomes 512 while the image size halves. The number of layers is 8 in this case (the input image size = 28) to have features of 1×1 dimension at the end of the encoder. The decoder reconstructs the output image of the same dimensions with the input from the acquired features of the encoder. The up-sampling blocks consist of ReLu, bilinear interpolation (factor of 2), 3 × 3 convolutions with stride 1 and pad 1, and batch normalization. Magnification with convolution was used instead of transposed convolution to avoid the checkerboard artifact [39]. The dimensions of blocks are symmetric with respect to the encoder, except that features of the same dimension in the encoding path are concatenated to preserve spatial information. The last blocks of the encoder and decoder do not contain the batch normalization.
Appendix B: Loss function
Among various metrics to compare two images, we utilized l1 norm [Eq. (1)] and SSIM (contrast component only) [Eq. (2)] [40]. We only exploited the contrast component because the global phase, which contributes to the offset, does not have any physical meaning. For network G, a network input x, and ground truth (GT) y,
When the loss function consisted of SSIM only, the optimization process was highly unstable, failing frequently. The performance of each loss combination was tested by calculating the average field cross-correlation error (FCE) [Eq. (4)] and phase RMSE [Eq. (5)]. For a network output and GT complex electric field E = Aeiϕ,
Appendix C: Dataset and training
Our network learned the five types of eukaryotic cells (HeLa, NIH-3T3, HEK-293 T, MDA-231, and COS7). The trained model can remove the aberration patterns in the same type of cells and can also be used for different types of data such as microspheres. To obtain the dataset, the optical field images of biological cells and microsphere were measured using a commercial 3D QPI instrument (HT-2H, Tomocube Inc., South Korea). The dataset was prepared by pairing the optical fields with aberrations and fields with aberration correction using the background subtraction method. To exploit the amplitude-phase correlation, the network was designed to handle 2-channel (amplitude and phase channel) images for input and output. For the ground truth training dataset, phase images were retrieved using a phase retrieval algorithm [41,42], and the 2π ambiguities in phase maps were resolved using a phase unwrapping algorithm, namely Goldstein’s algorithm [43].
After we split the acquired dataset into 1:1:1 for training, validation, and test, we trained our deep neural network with a loss function consisting of structural similarity index map and L1 norm [40]. We used the ADAM optimizer with an initial learning rate of 0.0002 and a beta range of (1, 0.999). We used a desktop computer with a CPU (XeonTM CPU E5-2630, Intel), and one graphic processing unit (GeForce GTX 1080Ti, 11 GB internal memory, NVIDIA) for training and testing. Once the training is completed, the process of converting an image with aberration into an aberration-corrected image takes 0.8 ms (with a batch size of 100) and 15 ms (with a single-image batch) using a single GPU. This non-linear time difference between two different batch sizes is caused by the GPU parallelization. All optical field images were cropped to 256 × 256 pixels before entering the network.
The learning curve of our network is depicted in Fig. 4(a). The average loss was calculated using the training set (11809 images) and validation set (11858 images) at each epoch during the training. In Fig. 4(b), 24-hour and 240-hour training results are presented. The results show that our network can be well optimized within 24 hours of training and validation. Nonetheless, we decided to train the network for sufficient time (approximately 240 hours) to fully learn the image-to-image translation. We stopped the training when the validation loss achieved 28.53 at 2223 epochs.
We also conducted a training with a smaller data size (Fig. 5). For comparison, we trained the new network with the same hyper-parameters as used in the original model for the same period of time. Smaller training data size does not fully leverage the learning capacity of our deep neural network. The inferred phase image is blurry, and the spherical aberration pattern is not clearly eliminated, compared to the original large data model. The data size is also correlated with a network architecture information capacity, and the result may be improved by optimizing the network as per the smaller data size. However, we believe that our training data size has been justified through this training with a small data model.
Appendix D: Demonstration on microbeads and various eukaryotic cells
To validate that the proposed network did not learn the specific sample distributions in the training dataset, but the ensemble of aberration patterns of the optical system used, we also imaged known polystyrene (PS) microsphere (Fig. 6). The phase images of the PS microsphere are presented in Fig. 6(a). The input phase image containing aberration and phase wrapping (left), the network output (middle), and the GT (right) obtained from the background subtraction method are visualized. Phase delay along the red dashed lines in the insets of Fig. 6(a) are also plotted in Fig. 6(b).
We also compared our method to the background subtraction on the performance of optical diffraction tomography in Fig. 6(c). After 49 angles of optical fields were obtained, each set of optical fields were processed with our method and the background subtraction, respectively. The bead tomograms reconstructed from two different sets of optical fields are shown: our method (left) and background subtraction (right). RI values along the blue dashed lines in Fig. 6(c) are plotted in Fig. 6(d). Although the network was trained using the dataset that consists of only the eukaryotic cells, our method has a close match with the background subtraction in both 2-D QPI and tomographic reconstruction. Also, we have applied the present algorithm to various types of cancer cell lines [Fig. 6(f)]. The QPI images of the HeLa, COS-7, and MDA-MB-213 cell were clearly retrieved with the present method.
Appendix E: Statistical analysis of results
The two types of errors (FCE and phase RMSE) are calculated for whole datasets (Fig. 7). For the images obtained from normal illumination only, 85% of the data in the test set has smaller FCE and phase RMSE than 0.028 and 0.15 rad, respectively. For the whole dataset containing images obtained from angled illumination, 85% of the data in the test set has smaller FCE and phase RMSE than 0.10 and 0.33 rad, respectively.
Appendix F: Temporal stability of our algorithm
To test the temporal stability of our approach, we inferred on data acquired later at various time points (Fig. 8). After completing the training, validation, and test using the dataset acquired in the years 2017-2018, we additionally tested on the dataset obtained later at four different time points in 2019. We confirmed that our trained model could successfully remove the aberration of the QPI data acquired even approximately six months later. Also, we computed the mean phase error of the inferred data, compared to the background subtraction method. While the data acquired at the time 2 and 3 seems to have a slightly larger error, the later time points 4 and 5 shows similar levels of error in comparison to the original data group 1, which implies the high temporal stability of our method.
Appendix G: Experiments on various data distributions
Appendix H: Optical setup and image reconstruction process
For 2D and 3D QPI, we used a commercialized imaging setup based on Mach-Zehnder interferometry with a high-speed illumination scanning using a digital micromirror device (DMD) (HT-2H, Tomocube Inc., South Korea) (Fig. 10). A laser beam from a 532-nm diode-pumped solid-state laser is split into a sample and a reference arm through a fiber coupler. The sample beam, passing through a lens 1 (L1), is modulated by a DMD (DLP6500FYE, Texas Instruments, USA) for angle scanning [5,44]. The beam is collected by a condenser lens (Obj 1, NA = 1.2, 60) to illuminate a sample with a controlled angle. The scattered light from the sample is projected to the image plane through a series of the lens (Obj 2, NA = 1.2, 60) and mirror. At the image plane, both the sample and reference beams interfere and form a spatially modulated interferogram, which is captured using an image sensor (FL3-U3-13Y3M-C, FLIR Systems, USA).
From each measured interferogram, the corresponding 2D optical field image is retrieved using a Fourier-transform-based phase retrieval algorithm [41,42]. Then, the retrieved field images are processed to reconstruct a 3D RI tomogram using the Fourier diffraction theorem with the Rytov approximation, which is widely used to solve the inverse light diffraction of a weakly scattering object [24,45].
Funding
National Research Foundation of Korea (2015R1A3A2066550, 2017M3C1A3013923, 2018K000396); KAIST (501100007107); Tomocube, Inc.
Acknowledgment
The authors appreciate Dr. Weisun Park (KAIST) and Dr. Sumin Lee (Tomocube Inc.) for providing biological samples.
Disclosures
Y. Jo, H.S. Min, G. Choi, and Y.K. Park have financial interests in Tomocube Inc., a company that commercializes ODT instruments and is one of the sponsors of the work.
References
1. Y. Park, C. Depeursinge, and G. Popescu, “Quantitative phase imaging in biomedicine,” Nat. Photonics 12(10), 578–589 (2018). [CrossRef]
2. R. W. Gerchberg, “A practical algorithm for the determination of phase from image and diffraction plane pictures,” Optik 35, 237 (1972).
3. G. Zheng, R. Horstmeyer, and C. Yang, “Wide-field, high-resolution Fourier ptychographic microscopy,” Nat. Photonics 7(9), 739–745 (2013). [CrossRef]
4. B. Kemper and G. von Bally, “Digital holographic microscopy for live cell applications and technical inspection,” Appl. Opt. 47(4), A52 (2008). [CrossRef]
5. K. Lee, K. Kim, G. Kim, S. Shin, and Y. Park, “Time-multiplexed structured illumination using a DMD for optical diffraction tomography,” Opt. Lett. 42(5), 999–1002 (2017). [CrossRef]
6. S. Kim, N. H. Kim, J. E. Park, J. W. Hwang, N. Myung, K.-T. Hwang, Y. A. Kim, C.-Y. Jang, and Y. K. Kim, “PRMT6-mediated H3R2me2a guides Aurora B to chromosome arms for proper chromosome segregation,” Nat. Commun. 11(1), 612 (2020). [CrossRef]
7. T.-K. Kim, B.-W. Lee, F. Fujii, K.-H. Lee, S. Lee, Y. Park, J. K. Kim, S.-W. Lee, and C.-G. Pack, “Mitotic Chromosomes in Live Cells Characterized Using High-Speed and Label-Free Optical Diffraction Tomography,” Cells 8(11), 1368 (2019). [CrossRef]
8. E. H. Hayakawa, K. Yamaguchi, M. Mori, and G. Nardone, “Real-time cholesterol sorting in Plasmodium falciparum-erythrocytes as revealed by 3D label-free imaging,” Sci. Rep. 10(1), 1–12 (2020). [CrossRef]
9. S. Cho, S. Kim, Y. Kim, and Y. Park, “Optical imaging techniques for the study of malaria,” Trends Biotechnol. 30(2), 71–79 (2012). [CrossRef]
10. J. Oh, J. S. Ryu, M. Lee, J. Jung, S. Han, H. J. Chung, and Y. Park, “Three-dimensional label-free observation of individual bacteria upon antibiotic treatment using optical diffraction tomography,” Biomed. Opt. Express 11(3), 1257–1267 (2020). [CrossRef]
11. Y. Jo, S. Park, J. Jung, J. Yoon, H. Joo, M.-h. Kim, S.-J. Kang, M. C. Choi, S. Y. Lee, and Y. Park, “Holographic deep learning for rapid optical screening of anthrax spores,” Sci. Adv. 3(8), e1700606 (2017). [CrossRef]
12. S. Park, J. W. Ahn, Y. Jo, H.-Y. Kang, H. J. Kim, Y. Cheon, J. W. Kim, Y. Park, S. Lee, and K. Park, “Label-Free Tomographic Imaging of Lipid Droplets in Foam Cells for Machine-Learning-Assisted Therapeutic Evaluation of Targeted Nanodrugs,” ACS nano (2020).
13. S. Shin, K. Kim, K. Lee, S. Lee, and Y. Park, “Effects of spatiotemporal coherence on interferometric microscopy,” Opt. Express 25(7), 8085–8097 (2017). [CrossRef]
14. F. Dubois, M. L. Requena, C. Minetti, O. Monnom, and E. Istasse, “Partial spatial coherence effects in digital holographic microscopy with a laser source,” Appl. Opt. 43(5), 1131–1139 (2004). [CrossRef]
15. Y. Park, W. Choi, Z. Yaqoob, R. Dasari, K. Badizadegan, and M. S. Feld, “Speckle-field digital holographic microscopy,” Opt. Express 17(15), 12285–12292 (2009). [CrossRef]
16. H. Farrokhi, J. Boonruangkan, B. J. Chun, T. M. Rohith, A. Mishra, H. T. Toh, H. S. Yoon, and Y.-J. Kim, “Speckle reduction in quantitative phase imaging by generating spatially incoherent laser field at electroactive optical diffusers,” Opt. Express 25(10), 10791–10800 (2017). [CrossRef]
17. B. Bhaduri, H. Pham, M. Mir, and G. Popescu, “Diffraction phase microscopy with white light,” Opt. Lett. 37(6), 1094–1096 (2012). [CrossRef]
18. Y. Baek, K. Lee, J. Yoon, K. Kim, and Y. Park, “White-light quantitative phase imaging unit,” Opt. Express 24(9), 9308–9315 (2016). [CrossRef]
19. P. Ferraro, S. De Nicola, A. Finizio, G. Coppola, S. Grilli, C. Magro, and G. Pierattini, “Compensation of the inherent wave front curvature in digital holographic coherent microscopy for quantitative phase-contrast imaging,” Appl. Opt. 42(11), 1938–1946 (2003). [CrossRef]
20. I. Choi, K. Lee, and Y. Park, “Compensation of aberration in quantitative phase imaging using lateral shifting and spiral phase integration,” Opt. Express 25(24), 30771–30779 (2017). [CrossRef]
21. W. J. Zhou, Y. J. Yu, and A. Asundi, “Study on aberration suppressing methods in digital micro-holography,” Opt. Laser. Eng. 47(2), 264–270 (2009). [CrossRef]
22. O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional Networks for Biomedical Image Segmentation,” in Med. Image Comput. Comput. Assist. Interv. – MICCAI 2015(Springer International Publishing, Cham, 2015), pp. 234–241.
23. P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation with conditional adversarial networks,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 2017) (2017), pp. 1125–1134.
24. E. Wolf, “Three-dimensional structure determination of semi-transparent objects from holographic data,” Opt. Commun. 1(4), 153–156 (1969). [CrossRef]
25. G. E. Spoorthi, S. Gorthi, and R. K. S. S. Gorthi, “PhaseNet: A Deep Convolutional Neural Network for Two-Dimensional Phase Unwrapping,” IEEE Signal Process. Lett. 26(1), 54–58 (2019). [CrossRef]
26. K. Wang, Y. Li, Q. Kemao, J. Di, and J. Zhao, “One-step robust deep learning phase unwrapping,” Opt. Express 27(10), 15100–15115 (2019). [CrossRef]
27. T. Nguyen, V. Bui, V. Lam, C. B. Raub, L. C. Chang, and G. Nehmetallah, “Automatic phase aberration compensation for digital holographic microscopy based on deep learning background detection,” Opt. Express 25(13), 15043–15057 (2017). [CrossRef]
28. J. Pan, S. Liu, D. Sun, J. Zhang, Y. Liu, J. Ren, Z. Li, J. Tang, H. Lu, Y. Tai, and M. Yang, “Learning Dual Convolutional Neural Networks for Low-Level Vision,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (IEEE, 2019) (2018), pp. 3070–3079.
29. M. Deng, S. Li, and G. Barbastathis, “Learning to synthesize: splitting and recombining low and high spatial frequencies for image recovery,” arXiv preprint, 1811.07945 (2018).
30. M. Ziemczonok, A. Kus, P. Wasylczyk, and M. Kujawinska, “3D-printed biological cell phantom for testing 3D quantitative phase imaging systems,” Sci. Rep. 9(1), 18872 (2019). [CrossRef]
31. A. Goy, K. Arthur, S. Li, and G. Barbastathis, “Low photon count phase retrieval using deep learning,” Phys. Rev. Lett. 121(24), 243902 (2018). [CrossRef]
32. Y. Rivenson, Y. Zhang, H. Günaydın, D. Teng, and A. Ozcan, “Phase recovery and holographic image reconstruction using deep learning in neural networks,” Light: Sci. Appl. 7(2), 17141 (2018). [CrossRef]
33. G. Zhang, T. Guan, Z. Shen, X. Wang, T. Hu, D. Wang, Y. He, and N. Xie, “Fast phase retrieval in off-axis digital holographic microscopy through deep learning,” Opt. Express 26(15), 19388–19405 (2018). [CrossRef]
34. G. Choi, D. Ryu, Y. Jo, Y. S. Kim, W. Park, H.-s. Min, and Y. Park, “Cycle-consistent deep learning approach to coherent noise reduction in optical diffraction tomography,” Opt. Express 27(4), 4927–4943 (2019). [CrossRef]
35. Y. Wu, Y. Luo, G. Chaudhari, Y. Rivenson, A. Calis, K. De Haan, and A. Ozcan, “Bright-field holography: cross-modality deep learning enables snapshot 3D imaging with bright-field contrast using a single hologram,” Light: Sci. Appl. 8(1), 25 (2019). [CrossRef]
36. Y. Jo, H. Cho, S. Y. Lee, G. Choi, G. Kim, H. S. Min, and Y. Park, “Quantitative Phase Imaging and Artificial Intelligence: A Review,” IEEE J. Sel. Top. Quantum Electron. 25(1), 1–14 (2019). [CrossRef]
37. A. L. Maas, A. Y. Hannun, and A. Y. Ng, “Rectifier nonlinearities improve neural network acoustic models,” in Proc. icml(2013), p. 3.
38. S. Ioffe and C. J. a. p. a. Szegedy, “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” arXiv preprint 03167 (2015).
39. A. Odena, V. Dumoulin, and C. Olah, “Deconvolution and checkerboard artifacts,” 1, e3 (2016).
40. Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Trans. Image Process. 13(4), 600–612 (2004). [CrossRef]
41. S. K. Debnath and Y. Park, “Real-time quantitative phase imaging with a spatial phase-shifting algorithm,” Opt. Lett. 36(23), 4677–4679 (2011). [CrossRef]
42. M. Takeda, H. Ina, and S. Kobayashi, “Fourier-Transform Method of Fringe-Pattern Analysis for Computer-Based Topography and Interferometry,” J. Opt. Soc. Am. 72(1), 156–160 (1982). [CrossRef]
43. R. M. Goldstein, H. A. Zebker, and C. L. Werner, “Satellite Radar Interferometry - Two-Dimensional Phase Unwrapping,” Radio Sci. 23(4), 713–720 (1988). [CrossRef]
44. S. Shin, K. Kim, J. Yoon, and Y. Park, “Active illumination using a digital micromirror device for quantitative phase imaging,” Opt. Lett. 40(22), 5407–5410 (2015). [CrossRef]
45. K. Kim, J. Yoon, S. Shin, S. Lee, S.-A. Yang, and Y. Park, “Optical diffraction tomography techniques for the study of cell pathophysiology,” J. Biomed. Photonics Eng. 2(2), 020201-1–020201-16 (2016). [CrossRef]