Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Coded-aperture broadband light field imaging using digital micromirror devices

Open Access Open Access

Abstract

Coded-aperture light field (CALF) imaging can record four-dimensional information of incident light rays with a high angular resolution while retaining a camera’s full pixel count. However, existing systems are limited by either imaging speeds and image contrasts of liquid-crystal spatial light modulators or by severe dispersion to broadband light of digital micromirror devices (DMDs), both of which hinder CALF’s widespread applications. Here, we overcome these limitations by developing dispersion-eliminated (DE) CALF imaging. Using a dual-DMD design to compensate for dispersion in the entire visible spectrum, the DECALF imaging system captures ${{1280}} \times {{1024}} \times {{5}} \times {{5}}$ $(x,y,\theta ,\varphi)$ color light field images at 20 Hz. Using static and dynamic three-dimensional (3D) color scenes, we experimentally demonstrate multi-perspective viewing, digital refocusing, and 3D tracking of the DECALF imaging system. We also apply it to the imaging and analyses of escape behaviors of freely moving normal and disease-model zebrafish.

© 2021 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

Light field imaging can record two-dimensional (2D) spatial $({x,{}y})$ and 2D angular $({\theta ,\varphi})$ information of incident rays [1]. The captured four-dimensional (4D) information endows light field imaging with attractive advantages, including multiple-perspective viewing, digital refocusing, and depth estimation [2]. To date, light field imaging has been widely implemented in microscopy [36], photography [7], and endoscopy [8,9]. In existing configurations, microlens arrays (MLAs) are popularly used to sample $({x,{}y})$ information in a field of view and fill in local voids with $({\theta ,\varphi})$ information [4]. Nonetheless, the induced trade-off poses challenges for MLA-based light field imaging to attaining high spatial resolution and high angular resolution simultaneously [10].

To overcome this problem, many efforts have been taken to capture 4D light fields with a camera’s full pixel count [1114]. Among these approaches, coded-aperture light field (CALF) imaging has sparked increasing research attention [1521]. This modality uses single or multiple masks to encode the system’s aperture. Despite retaining a camera’s full pixel count, early CALF imaging systems had various limitations, including low pattern adaptability to scenes [22], long acquisition time, and additional error due to pattern misalignment [16].

To improve the flexibility, efficiency, and accuracy in CALF imaging, liquid-crystal spatial light modulators (LC-SLMs) have been implemented for aperture encoding [1619]. Without any mechanically moving parts, LC-SLMs eliminate the error from mask misalignment. However, these systems suffer major drawbacks in contrast (due to imperfect polarization selectivity [23]), stability (due to the flicker noise [24]), and speeds (due to LC’s limited responsible time [19,25]). Thus far, CALF imaging of dynamic scenes at video rate is rarely performed.

Digital micromirror devices (DMDs) are a promising candidate to solve these problems [26,27]. As a 2D binary amplitude SLM [28], a DMD consists of up to millions of micromirrors, each of which can be independently tilted to either ${+}{{12}}^\circ$ or ${-}{{12}}^\circ$ from its surface normal to reflect incident light to one of the two directions as an “ON” or “OFF” pixel. This operating principle enables DMDs to produce high-contrast binary images. As a micro-electromechanical device, a DMD can generate binary patterns at up to tens of kilohertz [29]. Leveraging these technical advantages, DMDs have been used in phase-space measurements [30] (i.e., the physical-optics equivalence of light field imaging [31]). By placing a DMD on the Fourier plane to rapidly create and scan sub-apertures, this system recorded light field images of three red light-emitting diodes (LEDs) at different depths. However, acting as a diffraction grating, the DMD induced severe spatial dispersion in the acquired images for broadband light [23]. Consequently, a color filter with a 10-nm bandwidth was inserted to lessen the dispersion problem [30]. Thus, the DMD-based CALF imaging could be demonstrated only with static objects using narrowband light.

To surmount these limitations, we have developed a new dispersion-eliminated (DE) CALF imaging system for broadband light field imaging at video rate, as shown systematically in Fig. 1. Design details are presented in Fig. S1. Broadband light from an object is imaged by a ${{4}}f$ imaging system consisting of lenses L1 and L2. DMD1 (Texas Instruments, Discovery 4100, ${{0}.{{7}}^{{\prime \prime}}}$ XGA), placed on the back focal plane of L1, spatially disperses the incident light so that a spectrally smeared image of the object is formed on the intermediate image plane. This image is relayed to the final image plane by another identical ${{4}}f$ imaging system consisting of lenses L3 and L4 and DMD2 (Texas Instruments, Discovery 1100, ${{0}.{{7}}^{{\prime \prime}}}$ XGA). Because the two ${{4}}f$ imaging systems are symmetrical about the intermediate image plane, DMD1-induced dispersion is compensated for. In this way, a clear image of the object is formed on a high-speed color camera (PCO, 1200 hs) placed at the final image plane. In operation, DMD1 divides the system’s aperture into ${{5}} \times {{5}}$ square sub-apertures, each of which contains ${{50}} \times {{50}}$ micromirrors and has a 50% overlap with adjacent ones. An all-OFF pattern is loaded onto DMD2. The camera is synchronized with DMD1. Overall, the DECALF system acquires ${{1280}} \times {{1024}} \times {{5}} \times {{5}}$ $({x,y,\theta ,\varphi})$ light fields at 20 Hz.

 figure: Fig. 1.

Fig. 1. Schematic of the dispersion-eliminated coded-aperture light field (DECALF) system. DMD1–DMD2, digital micromirror devices; L1–L4, lenses (Thorlabs, AC254-100-A); M1–M2, mirrors (Thorlabs, PF10-03-G01).

Download Full Size | PDF

The characterization of the DECALF system was carried out by imaging a negative resolution target illuminated by a white LED (Thorlabs, MNWHL4) with a 400 nm–700-nm spectrum. An all-OFF pattern was loaded onto DMD1. The image captured on the intermediate image plane [Fig. 2(a)] shows severe dispersion induced by DMD1. In contrast, a clear image of the resolution target was captured at the final image plane [Fig. 2(b)], demonstrating that the dispersion is compensated for. The minimum resolvable feature sizes were quantified as 22.10 µm (group 4, element 4) in the horizontal direction [Fig. 2(c)] and 19.69 µm (group 4, element 5) in the vertical direction [Fig. 2(d)], both of which agree well with the theoretical values. The slight difference in the two directions is likely attributed to the unmatched surface curves of the two DMDs [32,33]. In addition, the axial resolution, depending on the camera’s pixel size and the system’s angular resolution [34], was determined to be 1.24 mm. Finally, the imaging volume, relying on the ($x,y$) field of view and the depth of field of perspective images, was quantified to be ${15.36} \times {12.29} \times {97.56}\;{\rm{mm}}^3$.

 figure: Fig. 2.

Fig. 2. Characterization of the DECALF system. (a) Image on the intermediate image plane. (b) Image on the final image plane. (c), (d) Averaged horizontal and vertical line profiles of selected elements on the resolution target [marked by red and blue lines in (b)]. Error bar: standard deviation.

Download Full Size | PDF

To demonstrate the DECALF system’s performance, we imaged a static three-dimensional (3D) color scene. As shown in Fig. 3(a), the incident white LED light was filtered by a multi-color filter (Izumar, Multi-color 58 mm). After that, a hollow maple-leaf mask and a “1X” symbol were placed at two depths separated by 64 mm. Perspective images were captured by sub-aperture scanning (see Visualization 1). Figure 3(b) shows four representative perspective images captured by opening the leftmost, rightmost, topmost, and bottommost sub-apertures, respectively. The first two panels illustrate the horizontal shift between the 1X symbol and the maple-leaf mask by opening two different sub-apertures along the horizontal direction. Similarly, the vertical shift is evident in the last two panels, corresponding to the opening of two sub-apertures along the vertical direction. Moreover, all perspective images retain the full pixel count of the deployed color camera. Using these perspective images, we digitally refocused the 3D scene to the front, to the back, and over the entire scene [Fig. 3(c)]. Full scanning of the digital refocus is provided in Visualization 2. The distance between the 1X symbol and the maple-leaf mask was quantified as 64.48 mm, which shows good agreement with the pre-set value.

 figure: Fig. 3.

Fig. 3. DECALF imaging of a static 3D color scene. (a) Experimental setup. (b) Four representative perspective images. (c) Digital refocusing results.

Download Full Size | PDF

To demonstrate the DECALF imaging’s ability to visualize dynamic objects, we imaged moving microspheres in water. In the experiment, the white LED illuminated polyethylene microspheres (Cospheric, WPMS-1.00 850–1000 µm) randomly distributed in water in a cuvette (Labshops, SKU:Q109), as shown in Fig. 4(a). The transmitted white light entered the DECALF system. Microspheres’ movement was induced by stirring the water. Figure 4(b) shows three all-focused images at 50 ms, 250 ms, and 400 ms, in which the depths of these microspheres (marked as M1–M5) were determined via digital refocusing. By calculating the centroids of each microsphere, time histories of 3D positions of these microspheres are plotted in Fig. 4(c), and the full evolution is shown in Visualization 3. In this experiment, the occlusion of microspheres was not observed. Nevertheless, it is worth noting that DECALF imaging is poised to mitigate such a problem with its current configuration. The acquired perspective images enable viewing the scene from different angles, which increases the chance to observe occluded microspheres. Using light field occlusion modeling [35], the depths of these microspheres could be estimated by the DECALF system.

 figure: Fig. 4.

Fig. 4. 3D tracking of moving microspheres using DECALF imaging. (a) Experiment setup. (b) Representative depth-coded images. (c) 3D positions of five microspheres over time.

Download Full Size | PDF

To highlight the dynamic 3D imaging ability of the DECALF system, we imaged six-day-old zebrafish larvae freely moving in a cuvette (Labshops, SKU:Q109) (Fig. 5, Fig S2, and Visualization 4). Water jetting was used to stimulate zebrafish’s escape behaviors. Three representative all-focused images of a zebrafish at 100 ms, 250 ms, and 600 ms are shown in Fig. 5(a). The time trace of the 3D spatial positions of the head of this zebrafish is shown in Fig. 5(b). Using this trajectory, we calculated its instantaneous moving velocities in the $x$, $y$, and $z$ directions [Fig. 5(c)]. To further analyze the zebrafish’s motion, we tracked the tail bending angle $\alpha$ and the fin orientation angle $\beta$. Changes in these angles, along with the zebrafish’s moving distance, are shown in Fig. 5(d). These results illustrate the correlation between the distance and the instantaneous velocities of the zebrafish. In addition, the results show that the tail bending angle is zero at the beginning and the end of the recording window, indicating that the zebrafish kept its tail straight when staying still. In contrast, once it encountered a threatening stimulus, large tail bending angles were observed, resulting in a change in direction followed by a rapid swim with higher instantaneous velocities. These behaviors are reflected in Fig. 5(c) as a sharp oscillation in its moving trace from 100 ms to 350 ms. Finally, the data reveal asymmetrical orientation angles of the left and right fins, indicating drastic changes in direction during the zebrafish’s escape from the stimulus.

 figure: Fig. 5.

Fig. 5. 3D tracking of a six-day-old freely moving zebrafish larva using the DECALF system. (a) Representative all-focused frames at 100 ms, 250 ms, and 600 ms. (b) 3D trace of the zebrafish. (c) Instantaneous moving velocities of the zebrafish in the $x$, $y$, and $z$ directions. (d) Time histories of the moving distance, tail bending angle, and fin orientation angle of the zebrafish.

Download Full Size | PDF

To test the DECALF imaging’s assessment of swimming behavioral differences in different zebrafish models, we applied it to imaging a normal zebrafish and a C9ORF72 loss-of-function (C9-LOF) zebrafish (Fig. 6 and Visualization 5). Recently developed to study the pathogenesis of amyotrophic lateral sclerosis, this C9-LOF zebrafish model replicates aspects of this disease, including motor behavioral defects, muscle atrophy, and motor neuron loss [36]. The representative all-focused frames of normal and C9-LOF six-day old zebrafish larvae at three time points [Fig. 6(a)] show no apparent difference in their shapes. However, when water stream stimulation was applied, we observed different behaviors by tracking the 3D positions of both models [Fig. 6(b)]. The normal zebrafish quickly moved away from the site of the startle. In contrast, the C9-LOF zebrafish showed slow responses and a limited moving ability due to motor deficits [36]. This difference is quantitatively reflected in the instantaneous velocities of the normal and C9-LOF zebrafish in the $x$, $y$, and $z$ directions, as shown in Fig. 6(c). While the curves for the normal zebrafish oscillate sharply in all three directions, those of the C9-LOF zebrafish show small changes, especially in the $x$ direction. Altogether, these findings demonstrate the power of DECALF imaging for the behavioral study of disease-model zebrafish in vivo.

 figure: Fig. 6.

Fig. 6. Comparison of escape behaviors of a normal and disease-model (C9-LOF) zebrafish. (a) Representative all-focused frames at 50 ms, 100 ms, and 200 ms. Depths are coded with colors. Backgrounds are subtracted for better display. (b) 3D traces after stimulation. (c) Instantaneous moving velocities in the $x$, $y$, and $z$ directions.

Download Full Size | PDF

In summary, we have developed DMD-based DECALF imaging for high-resolution, color light field acquisition using broadband visible light at video rate. DECALF imaging is applied to studying zebrafish’s motion under stimulation. Circumventing the trade-off between spatial and angular resolutions, DECALF imaging enables ${{5}} \times {{5}}$ $({\theta ,\varphi})$ perspectives at the camera’s full $({x,y})$ pixel count of ${{1280}} \times {{1024}}$. DECALF imaging extends the operation scope of DMD-based CALF imaging to broadband light. Compared with conventional CALF imaging that employs a single DMD with a narrow-bandpass filter [30] or monochromatic illumination [23], the accommodation of the full visible spectrum in the DECALF system enhances light throughput. Its dispersion-compensated design also avoids the reduction of spatial resolution by pixel binning and the decrease in image quality due to laser speckles. Furthermore, the broadband imaging circumvents the potential color-induced complexity in the study of animal behaviors [37]. As a universal imaging scheme, the DECALF system is expected to be integrated into a variety of modalities for both macroscopic and microscopic light field imaging.

Our future work will focus on further improving the DECALF system’s technical specifications and imaging capability. For example, the 500-Hz full frame rate of the camera used in this work is much lower than the 22-kHz refreshing rate of the DMD. Replacing it with a high-speed camera [38,39] could largely increase the frame rate of light field imaging. Moreover, other advanced encoding schemes [4042] could be employed to considerably enhance the signal-to-noise ratio and accuracy in both the acquired perspective images and the recovered light field images. Finally, by enlarging the angular range covered by the perspective images and by employing super-resolution algorithms in digital refocusing [43], the DECALF system will enable accurate depth sensing in the scenario of partial occlusion, shedding new light on in vivo high-speed 3D position tracking.

Funding

Natural Sciences and Engineering Research Council of Canada (ALLRP 551076-20, ALLRP-549833-2020, CRDPJ-532304-18, RGPAS-507845-2017, RGPIN-2017-05959); Fonds de Recherche du Québec - Santé (267406, 280229); Canada Foundation for Innovation (37146); Fonds de recherche du Québec–Nature et technologies (2019-NC-252960).

Acknowledgment

The authors thank Prof. Michael F. Becker for providing DMD equipment and software, Qianwan Yang for experimental assistance, and Zoe Butti for assistance in breeding C9-LOF zebrafish. Shunmoogum A. Patten acknowledges the support from the Anna Sforza Djoukhajian Research Chair in ALS.

Disclosures

The authors declare no conflicts of interest.

Supplemental document

See Supplement 1 for supporting content.

REFERENCES

1. M. Levoy, Computer 39, 46 (2006). [CrossRef]  

2. M. Levoy, B. Chen, V. Vaish, M. Horowitz, I. McDowall, and M. Bolas, ACM Trans. Graph. 23, 825 (2004). [CrossRef]  

3. M. Levoy, R. Ng, A. Adams, M. Footer, and M. Horowitz, ACM Trans. Graph. 25, 924 (2006). [CrossRef]  

4. R. Prevedel, Y.-G. Yoon, M. Hoffmann, N. Pak, G. Wetzstein, S. Kato, T. Schrödel, R. Raskar, M. Zimmer, and E. S. Boyden, Nat. Methods 11, 727 (2014). [CrossRef]  

5. N. C. Pégard, H.-Y. Liu, N. Antipa, M. Gerlock, H. Adesnik, and L. Waller, Optica 3, 517 (2016). [CrossRef]  

6. Q. Geng, Z. Fu, and S.-C. Chen, J. Biomed. Opt. 25, 106502 (2020). [CrossRef]  

7. R. Ng, M. Levoy, M. Brédif, G. Duval, M. Horowitz, and P. Hanrahan, Stanford Technical ReportCTSR 2005-02 (2005), pp. 1–11.

8. J. Liu, D. Claus, T. Xu, T. Kessner, A. Herkommer, and W. Osten, Opt. Lett. 42, 1804 (2017). [CrossRef]  

9. E. Kwan, Y. Qin, and H. Hua, OSA Contin. 3, 194 (2020). [CrossRef]  

10. J. Chang, I. Kauvar, X. Hu, and G. Wetzstein, IEEE Computer Society Conference on Computer Vision and Pattern Recognition (2016), p. 3737.

11. X. Lin, J. Wu, G. Zheng, and Q. Dai, Biomed. Opt. Express 6, 3179 (2015). [CrossRef]  

12. J. Liu, T. Xu, W. Yue, J. Sun, and G. Situ, Opt. Express 23, 29154 (2015). [CrossRef]  

13. J. Unger, A. Wenger, T. Hawkins, A. Gardner, and P. Debevec, Eurographics Workshop on Rendering (2003), p. 141.

14. A. Veeraraghavan, R. Raskar, A. Agrawal, A. Mohan, and J. Tumblin, ACM Trans. Graph. 26, 69 (2007). [CrossRef]  

15. S. D. Babacan, R. Ansorge, M. Luessi, P. R. Mataran, R. Molina, and A. K. Katsaggelos, IEEE Trans. Image Process. 21, 4746 (2012). [CrossRef]  

16. C.-K. Liang, T.-H. Lin, B.-Y. Wong, C. Liu, and H. H. Chen, ACM Trans. Graph. 27, 1 (2008). [CrossRef]  

17. H. Y. Liu, J. Zhong, and L. Waller, Opt. Express 25, 14986 (2017). [CrossRef]  

18. H. Nagahara, C. Zhou, T. Watanabe, H. Ishiguro, and S. K. Nayar, IPSJ Trans. Comput. Vis. Appl. 4, 1 (2012). [CrossRef]  

19. C. Zuo, J. Sun, S. Feng, M. Zhang, and Q. Chen, Opt. Lasers Eng. 80, 24 (2016). [CrossRef]  

20. J. Liang, Rep. Prog. Phys. 83, 116101 (2020). [CrossRef]  

21. K. Marwah, G. Wetzstein, Y. Bando, and R. Raskar, ACM Trans. Graph. 32, 1 (2013). [CrossRef]  

22. X. Wang, F. Dai, Y. Ma, K. Gao, and Y. D. Zhang, Multimed. Tools Appl. 78, 697 (2019). [CrossRef]  

23. H.-Y. Liu, “Optical phase space measurements and applications to 3D imaging and light scattering,” Ph.D. thesis (University of California, 2018).

24. R. F. Voss, 33rd Annual Symposium on Frequency Control (1979), p. 40.

25. Z. Zhang, Z. You, and D. Chu, Light Sci. Appl. 3, e213 (2014). [CrossRef]  

26. A. Chiranjan, B. Duvenhage, and F. Nicolls, Pattern Recognition Association of South Africa and Robotics and Mechatronics International Conference (PRASA-RobMech) (2016), p. 1.

27. M. Hoffmann, I. N. Papadopoulos, and B. Judkewitz, Opt. Lett. 43, 22 (2018). [CrossRef]  

28. J. Liang, M. F. Becker, R. N. Kohn, and D. J. Heinzen, J. Micro. Nanolithogr. MEMS MOEMS 11, 023002 (2012). [CrossRef]  

29. J. Liang, S.-Y. Wu, R. N. Kohn, M. F. Becker, and D. J. Heinzen, Opt. Eng. 51, 108201 (2012). [CrossRef]  

30. H.-Y. Liu, E. Jonas, L. Tian, J. Zhong, B. Recht, and L. Waller, Opt. Express 23, 14461 (2015). [CrossRef]  

31. L. Waller, G. Situ, and J. W. Fleischer, Nat. Photonics 6, 474 (2012). [CrossRef]  

32. V. Parthiban, R. N. Kohn Jr., J. Liang, and M. F. Becker, Proc. SPIE 9761, 97610M (2016). [CrossRef]  

33. C. Gong and T. Hogan, IEEE J. Electron Devices Soc. 2, 27 (2014). [CrossRef]  

34. M. Martínez-Corral and B. Javidi, Adv. Opt. Photon. 10, 512 (2018). [CrossRef]  

35. T.-C. Wang, A. A. Efros, and R. Ramamoorthi, IEEE Trans. Pattern Anal. Mach. Intell. 38, 2170 (2016). [CrossRef]  

36. Z. Butti, J. Giacomotto, and S. Patten, Preprint (Version 1) available at Research Square https://doi.org/10.21203/rs.3.rs-49118/v1 (2020).

37. G. A. Mohamed, R.-K. Cheng, J. Ho, S. Krishnan, F. Mohammad, A. Claridge-Chang, and S. Jesuthasan, BMC Biol. 15, 103 (2017). [CrossRef]  

38. T. G. Etoh, T. Okinaka, Y. Takano, K. Takehara, H. Nakano, K. Shimonomura, T. Ando, N. Ngo, Y. Kamakura, and V. T. S. Dao, Sensors 19, 2247 (2019). [CrossRef]  

39. X. Liu, J. Liu, C. Jiang, F. Vetrone, and J. Liang, Opt. Lett. 44, 1387 (2019). [CrossRef]  

40. Y. Inagaki, Y. Kobayashi, K. Takahashi, T. Fujii, and H. Nagahara, European Conference on Computer Vision (2018), p. 418.

41. J. Liang, L. Gao, C. Li, and L. V. Wang, Opt. Lett. 39, 430 (2014). [CrossRef]  

42. L. Tian, X. Li, K. Ramchandran, and L. Waller, Biomed. Opt. Express 5, 2376 (2014). [CrossRef]  

43. V. Boominathan, K. Mitra, and A. Veeraraghavan, International Conference on Intelligent Computer Communication and Processing (ICCP) (2014), p. 1.

Supplementary Material (6)

NameDescription
Supplement 1       DECALF Supplementary Materials
Visualization 1       Perspective images of a static 3D color scene captured by the DECALF system.
Visualization 2       Digital refocusing of a static 3D color scene in DECALF imaging.
Visualization 3       3D position tracking of moving polyethylene microspheres.
Visualization 4       3D position tracking of a freely moving zebrafish.
Visualization 5       3D position tracking of a normal and a C9-LOF zebrafish.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (6)

Fig. 1.
Fig. 1. Schematic of the dispersion-eliminated coded-aperture light field (DECALF) system. DMD1–DMD2, digital micromirror devices; L1–L4, lenses (Thorlabs, AC254-100-A); M1–M2, mirrors (Thorlabs, PF10-03-G01).
Fig. 2.
Fig. 2. Characterization of the DECALF system. (a) Image on the intermediate image plane. (b) Image on the final image plane. (c), (d) Averaged horizontal and vertical line profiles of selected elements on the resolution target [marked by red and blue lines in (b)]. Error bar: standard deviation.
Fig. 3.
Fig. 3. DECALF imaging of a static 3D color scene. (a) Experimental setup. (b) Four representative perspective images. (c) Digital refocusing results.
Fig. 4.
Fig. 4. 3D tracking of moving microspheres using DECALF imaging. (a) Experiment setup. (b) Representative depth-coded images. (c) 3D positions of five microspheres over time.
Fig. 5.
Fig. 5. 3D tracking of a six-day-old freely moving zebrafish larva using the DECALF system. (a) Representative all-focused frames at 100 ms, 250 ms, and 600 ms. (b) 3D trace of the zebrafish. (c) Instantaneous moving velocities of the zebrafish in the $x$, $y$, and $z$ directions. (d) Time histories of the moving distance, tail bending angle, and fin orientation angle of the zebrafish.
Fig. 6.
Fig. 6. Comparison of escape behaviors of a normal and disease-model (C9-LOF) zebrafish. (a) Representative all-focused frames at 50 ms, 100 ms, and 200 ms. Depths are coded with colors. Backgrounds are subtracted for better display. (b) 3D traces after stimulation. (c) Instantaneous moving velocities in the $x$, $y$, and $z$ directions.
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.