Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Lensfree diffractive tomography for the imaging of 3D cell cultures

Open Access Open Access

Abstract

New microscopes are needed to help realize the full potential of 3D organoid culture studies. In order to image large volumes of 3D organoid cultures while preserving the ability to catch every single cell, we propose a new imaging platform based on lensfree microscopy. We have built a lensfree diffractive tomography setup performing multi-angle acquisitions of 3D organoid culture embedded in Matrigel and developed a dedicated 3D holographic reconstruction algorithm based on the Fourier diffraction theorem. With this new imaging platform, we have been able to reconstruct a 3D volume as large as 21.5 mm3 of a 3D organoid culture of prostatic RWPE1 cells showing the ability of these cells to assemble in 3D intricate cellular network at the mesoscopic scale. Importantly, comparisons with 2D images show that it is possible to resolve single cells isolated from the main cellular structure with our lensfree diffractive tomography setup.

© 2016 Optical Society of America

1. Introduction

The study of in vitro cell populations remains a challenging task if one needs to gather large quantitative and systematic data over extended period of time while preserving the integrity of the living sample. As discussed in [1], there is a need for a new microscopy technique that must be label-free and non-phototoxic to be as “gentle” as possible with the sample, and “smart” enough to observe the sample exhaustively at a variety of scales both in space and time. Lensfree video microscopy is addressing these needs in the context of 2D cell culture [2–4].

As scientists better understand the benefit of growing organoids in 3D and routinely adopt 3D culture techniques, lensfree imaging must also be adapted to 3D cultures. Therefore, the new challenging task is to extend lensfree microscopy techniques to the acquisitions and 3D reconstructions of large organoids structures [5–7]. The adaptation of lensfree microscopy techniques to 3D organoid cultures imaging is the scope of the present paper.

We first describe an experimental bench dedicated to lensfree diffractive tomography of 3D biological samples. Next, we present a 3D reconstruction algorithm based on diffractive tomographic microscopy methods [8–11] but applied here to lensfree microscopy. And as a first proof of principle, we present the 3D reconstruction of a RWPE1 prostatic cell culture in 3D within an extra-cellular matrix (Matrigel). The reconstructed volume is as large as 21.5 mm3 which allows to observe simultaneously a very large cellular networks and isolated cells.

2. Materials and methods

2.1. Cell lines and 3D organoid culture

RWPE1 (ATCC CRL-11609) cells were derived from non-neoplastic human prostate epithelial cells immortalized with human papillomavirus. Cells were maintained in KSFM (Life Technologies, Carlsbad, CA, ref. 17005-075) supplemented with 5 ng/mL epidermal growth factor (EGF) and 50 µg/mL bovine pituitary extract. The cells were maintained in culture until approximately 70% confluency. For passaging, cells were washed with Dulbeccos Ca2+/Mg-free PBS (D-PBS, Life Technologies, ref. 14190) and incubated with 1 mL trypsin-EDTA (Lonza, Basel, CH, ref. CC-5012, 0.25 mg/mL) for approximately 7 minutes. The trypsin was neutralized with 2 mL trypsin neutralizing solution (Lonza, ref. CC-5002), and the cells were recovered by centrifugation and counted using a Scepter 2.0 Handheld Automated Cell Counter (Millipore, Billerica, MA, ref. PHCC 20060). The RWPE1 cells were cultured in 3D with KSFM (Life Technologies, ref. 17005-075) supplemented with 50 ng/mL EGF and 2% fetal bovine serum (FBS). The 3D culture was grown in Matrigel (BD Biosciences, San Jose, CA, ref. 356231) according to a no top-coat protocol. Briefly, Matrigel was thawed overnight and poured into 4-well (160 µL of Matrigel) or 8-well Labtek® (90 µL of Matrigel®) plates on ice. For polymerization, Matrigel was incubated for 30 minutes at 37°C. Cells were seeded in half the final volume (500 µL or 250 µL of culture media in 4-well or 8-well Labtek respectively) and allowed to adhere for approximately 45 minutes. Then culture media was slowly poured over the attached cells and was changed every other day. All cells were routinely cultured in a humidified atmosphere with 5% CO2 at 37°C.

2.2. Experimental bench

Unlike 2D lensfree imaging, where only one image is required for retrieving the 2D object, the reconstruction of a 3D object from lensfree acquisitions requires to multiply the viewing angles. For this purpose, we have developed an experimental bench, illustrated on Fig. 1. It is composed of a semi-coherent illumination source (LED CREE, λ = 520 nm - ref. XLamp MCE RGBW MCE4CT) and CMOS sensor (IDS - 29.4 mm2, 3840 × 2748 monochromatic pixels, pixel pitch 1.67 µm - ref. UI-1942LE-M).

 figure: Fig. 1

Fig. 1 Left-hand side - Experimental bench dedicated to lensfree diffractive tomography. Right-hand side - Optical schema of the system. The semi-coherent incident plane wave Uinc passes through the sample volume. Each element of the volume diffracts the incident plane, behaving like secondary spherical sources, creating a diffracted wave Udif. The sensor records the intensity of their summation: Itot = |Utot|2 = |Uinc + Udif|2

Download Full Size | PDF

The object and the sensor are static, the object being placed on top of the sensor at a distance of 1 to 3 mm. The source is rotated with a rotation stage (Newport - ref. M-URM80APP, controlled by a Newport universal motion controller - ref. ESP300) about the φ-axis (see Fig. 1 and 2) and thus the illumination direction is tilted relative to the sensor plane. This configuration is the most adapted to 3D organoid culture in standard containers such as Petri dish or well plates.

2.3. Reconstruction methods

In the case of 3D lensfree imaging, the modelling of the wave Udif diffracted by the sample (see Fig. 1), defined in each point r=(x,y,z) of the space, is given by [8]:

Udif(r)=14πObjectf(r)Uinc(r)exp(ik0rr)rrd3r=14π(f.Uinc)*h
where k0=2πn0λ0 is the wave number in an ambient medium of refractive index n0 and h(r)=1rexp(ik0r) is the convolution kernel characteristic of spherical wave propagation. The scattering potential f(r) is the 3D object to be reconstructed. It is related to the complex refractive index of the object n(r) by the formula:
f(r)=k02((n(r)n0)21)

These equations are valid when the overall phase delay introduced by the sample is weak (see [9]). In the case of 3D cell culture, this requirement is generally not met since these samples feature objects with dimensions larger than l = 20 µm. In water, for a typical cell refractive index of 1.35, the difference of refractive index is δn = 0.02, introducing a phase delay of 2πlδn/λ. In visible light (λ ≃ 550 nm), this phase delay is greater than π and therefore cannot be supposed to be weak. In the present work, we decided to use these equations nonetheless to determine if some information, interesting on the biological point of view such as positions, dimensions and shapes of the objects could be estimated. As a drawback of the methodology, it is clear that quantitative results such as refractive indexes or optical lengths cannot be obtained.

2.3.1. Algorithm

Our reconstruction algorithm recovers f based on N multiple acquisitions at different angles (φj)j∈⟦1,N (see Fig. 2), and is based on the Fourier diffraction theorem, widely used in the field of diffractive tomographic microscopy [9–11]. This latter links U^dif, the 2D Fourier transform of Udif to f^, the 3D Fourier transform of the object f. At a given illumination angle φj on the sensor plane z = zs, Eq. (1) becomes (see [8, 12]):

f^(αj,βj,γj)=4iπwexp(2iπwzs)Hzs{k0j;λ0}U^difj(u,v;zs)
with, {αj=un0p0jλ0βj=vn0q0jλ0γj=wn0m0jλ0 and w=n02λ02u2v2, where (u, v) and (α, β, γ) correspond respectively to the Fourier space coordinates of the sensor and the object. (p0j,q0j,m0j) is the orientation of the φj illumination, i.e. the orientation of the wave vector k0j:
{p0j=0q0j=sinφjm0j=cosφj

 figure: Fig. 2

Fig. 2 Schema of three different acquisitions at the illumination angles φj1, φj2 and φj3, with (j1, j2, j3) ∈ ⟦1, N3, where N is the total number of acquisitions.

Download Full Size | PDF

Notice that the Fourier transform is defined for a given function g as:

g^(u)=g(x)exp(2iπux)dx

Figure 3 illustrates the Fourier diffraction theorem: a geometrical description is that the 2D Fourier transform U^difj of a given diffracted wave Udifj is mapped on a spherical cap (2D surface, also called Ewald’s sphere) depending on (k0j;λ0) inside the 3D frequency space of the object of interest f. To actually map the uniform 3D discrete Fourier space of f^, the discrete Fourier transform of each projection is computed. The position of each calculated 2D Fourier coefficient on its spherical cap is then determined and its value is associated to the nearest neighbour into this uniform frequency volume. Once all the 2D Fourier transforms of projections are mapped, it results to gaps within the 3D frequency space, as it can be seen in the medallion of Fig. 3. In this paper, it was decided not to apply any interpolation method to generate a fully filled 3D frequency support. Finally a 3D inverse discrete Fourier transform retrieves the object f.

 figure: Fig. 3

Fig. 3 Illustration of the Fourier diffraction theorem with the notations of Fig. 2. The 3D frequency space of the object of interest f is mapped with the 2D Fourier transform of the projections Udifj1, Udifj2 and Udifj3. An example of a Fourier region which is actually filled by the algorithm is given in the medallion in the 3D frequency domain part of the figure with the parameters used in section 3 to reconstruct a given region of interest: a volume of 1.34 × 1.34 × 1.34 mm3 with a voxel size of 1.67 × 1.67 × 3.34 µm3 at λ = 520 nm with an illumination angle varying from −30° (blue) to 30° (green) with a step angle of 5°. The red cap corresponds to the region map with the normal illumination.

Download Full Size | PDF

In sum, the algorithm operates as follows:

f=3D1[S(Hzsj2D[M0jUdifj])|{k0j|j=1N}]
where f constitutes the 3D image of size Nx × Ny ×Nz to be reconstructed. 3D1 and 2D are respectively the inverse 3D and direct 2D discrete Fourier transforms. They are performed by the FFT algorithm. S stands for the mapper into the 3D frequency space of the Nj spherical caps corresponding to Nj projections Udifj to which a given illumination wave vector k0j is associated. The projections are 2D complex images of size Ndx×Ndy corresponding to the sensor’s size, with a pixel pitch denoted Δd, which is the same in each direction (see Fig. 1). M0j is the modulation of Udifj in the spatial domain corresponding to the translation of the spherical cap from the sensor’s coordinate space (u, v) to the object’s coordinate space (αj, βj, γj) in the frequency domain in Eq. (3) (see Fig. 3):
M0j(x,y)=exp(2iπ(n0p0jλ0x+n0q0jλ0y))

2.3.2. An approximation for (Udifj)j1,N

In our case, only the intensity of the total transmitted wave Itotj=|Utotj|2 is acquired by the sensor, where Utot = Uinc + Udif (see Fig. 1 and 2). Hence we have to approximate the diffracted wave Udifj from Itotj. To do so, we first normalize Itotj so that the background value, corresponding to the unperturbed incident wave, is equal to 1. We note this normalized projection I¯totj. We multiply the square root of I¯totj by a phase ramp exp(ik0jr) to take into account the non-uniform phase shift induced by the tilted incidence of k0j relative to the sensor plane. Then we subtract the theoretical incident plane wave Uincj=exp(ik0jr):

UdifjI¯totjexp(ik0jr)Uincj

This formula is schematically represented on Fig. 4. This is a needed step in the presented methods. Indeed, the phase is very important since, inter alia, it contains the inclination of the wavefront. Without this information, the mapping of the Fourier space with Eq. (5) cannot work because the computed 2D spectra will be shift in the Fourier space. Equation (7) allows to reintroduce the information of the illumination angle in the data even if the simulated phase is just an approximation of the real one.

 figure: Fig. 4

Fig. 4 Illustration of the steps to get an approximation of Udifj for a given j.

Download Full Size | PDF

This calculation does not totally compensate the lack of phase information at the sensor plane. In particular, the phase distortion introduced by the object is not taken into account and could be approximated by means of a phase retrieval algorithm in order to diminish the artefactual twin image [13].

2.3.3. Reconstruction of regions of interest

If we take 1.67×1.67×1.67 µm3 voxels based on the pixel pitch of the sensor (see Fig. 1), the size of the 3D reconstructed image is very huge - up to several tens of gigabytes - if we want to maintain the same resolution in each direction ( NxNdx, NyNdy and NzNdx or Ndy). This could be too expensive for conventional computers and could cause memory overflows. To address this problem, we can extract from the data a region of interest (ROI) and reconstruct it separately. The extraction process of a ROI is illustrated on Fig. 5. The ROI is defined on the projection corresponding to φ = 0°, i.e. where the incidence of illumination is normal to the sensor. Knowing the altitude zs and the angle φj, we can predict where the particular structures in the ROI have been shifted in the other projections, due to the tilted incidence. The shift is theoretically known and equals:

{x0j=0y0j=zstanφj

 figure: Fig. 5

Fig. 5 3D lensfree data of a culture of prostatic cells RWPE1 in Matrigel. 61 angles of view were acquired from −30° to 30° with an angular pitch of 1° around the x axis. The illumination wavelength is 520 nm. A region of interest (ROI) is extracted on each projection and centered by registration. This aims at performing a localized reconstruction to save computation from RAM overflows. The centering has to be taken into account in the algorithm (Eq. (5)).

Download Full Size | PDF

The centering of the ROI whatever the angle of view implies to modify the algorithm to indicate the absolute position of the ROI on each projection. As it is a translation in the spatial domain, this yields to a modulation of the Fourier transform of the projection as a function of (x0j,y0j), giving from Eq. (5):

f=3D1[S(HzsjD0j2D[M0jUdifROIj])|{k0j,x0j,y0j|j=1Nj}]
where
D0j(u,v)=exp(2iπ(ux0j+vy0j))
and
M0j(x,y)=exp(2iπ(n0p0jλ0(x+x0j)+n0q0jλ0(y+y0j))).

As a first approximation, we consider that the emitted light propagates in a homogeneous medium. As a result, it follows a straight path until it is diffracted by the objects of interest. However, our setup induces several refractive indexes’ changes (air, culture medium, glass) along the light path. This causes uncertainties on the effective shift (x0j,y0j) as a function of the illumination angle φj. Moreover, the altitude zs is not known absolutely and has to be estimated with a focusing algorithm using Fresnel backpropagation at incidence φ = 0°. As a solution, we use a registration algorithm to correctly center the extracted ROIs so that the projections match the theoretical shift. We only consider rigid translations, ignoring the fact that the holograms are also stretched differently as a function of the light incidence.

3. Results

We have acquired tomographic data of a culture of prostatic cells RWPE1 in Matrigel, over 61 angles from −30° to 30° with an angular pitch of 1°. Figure 5 shows 3 projections at φ = 30°, φ = 0° and φ = 30°. We have reconstructed a large volume of 4 × 4 × 1.34 mm3 showing that the cells interestingly tend to form a structured network.

To reconstruct such a large volume, we worked by pieces of 1.34 × 1.34 × 1.34 mm3, i.e. nine reconstructions. Thus we have divided the data into nine adjacent ROIs of size 1.34 × 1.34 mm2 (800 × 800 pixels), as illustrated on Fig. 6. The registration algorithm for the centering of each projection was applied relative to the central ROI (dash red on Fig. 6), as well as the focusing algorithm to estimate zs, the distance between the sensor plane and the cell culture. The eight other ROIs remain adjacent to this one whatever the angle of view, to maintain the continuity in the 3D reconstruction. Each reconstructed 3D image had a size of 800 × 800 × 400 voxels with the sampling rate 1.67 × 1.67 × 3.34 µm3. Owing this decoupling task, a single ROI reconstruction has a size of two gigabytes in double precision which can be handled on a desktop computer. As a result the global 3D image has a size of 2400 × 2400 × 400 voxels with the same sampling rate, this whole image having a size close to 20 gigabytes.

 figure: Fig. 6

Fig. 6 Extraction of 9 ROIs of size 1.34 × 1.34 mm2 (800 × 800 pixels) for the piecewise 3D reconstruction of the 3D culture of prostatic cells RWPE1 presented on Fig. 5.

Download Full Size | PDF

The whole reconstructed volume f is shown on Fig. 7. As it is a complex image, we only visualize its imaginary part which gave us the most observable information. Three different sectional views - xy (center), xz (right) and yz (bottom) - and 3D views (see Visualization 1 and Visualization 2) show that the network formed by the cells is well focused at precise altitudes in the volume. The network fits a surface which is not horizontal, which demonstrates that the reconstruction algorithm effectively operates in 3D: isolated objects are also focused apart from the network at independent altitudes. Some of them are pointed out by red arrows on Fig. 7.

 figure: Fig. 7

Fig. 7 Piecewise 3D reconstruction (imaginary part of f) of a large volume (4 × 4 × 1.34 mm3) of the culture of prostatic cells RWPE1 from tomographic data presented on Fig. 5. (a) 3 sectional views - xy (center), xz (right) and yz (bottom) - are shown, each one focused at a given distance d = zs − z from the sensor. (b) 3D views of the reconstruction from different observation points (see Visualization 1 and Visualization 2). The dash red curves indicate the tilt of the network relative to the horizontal plane. The red arrows point out isolated objects that are not focused at the same altitude as the network.

Download Full Size | PDF

The central ROI and the ROI in the top-right corner of the reconstructed volume shown on Fig. 7 (equivalently the central and bottom-right ROIs on the raw data on Fig. 6) and are further detailed on Fig. 8 and 9. Each of these figures represents the three sectional views - xy (center), xz (right) and yz (bottom) - and a 3D view of the ROI and two specific patterns such as a cell or an organoid which are pointed by arrows in the ROIs. The rings around the focused objects on the xy views are artefactual: this is the well-known artefactual phenomenon of twin image, induced by the lack of absolute phase information in the data.

 figure: Fig. 8

Fig. 8 ROI at the center of the piecewise 3D reconstruction of the volume of Fig. 7. (a) 3 sectional views - xy (center), xz (right) and yz (bottom) - and a 3D view of the ROI. (a) same representation rules on two specific organoids at the full resolution of the reconstruction (see Visualization 3, Visualization 4, and Visualization 5). These are pointed by arrows on the global view. The dashed-line indicates the network plane. The red arrow points toward a 33 × 38 × 42 µm3 object and the blue arrow toward a 16 × 17 × 40 µm3 object.

Download Full Size | PDF

 figure: Fig. 9

Fig. 9 ROI at the to right corner of the piecewise 3D reconstruction of the volume of Fig. 7. (a) 3 sectional views - xy (center), xz (right) and yz (bottom) - and a 3D view of the ROI. (b) same representation rules on two specific organoids at the full resolution of the reconstruction (see Visualization 6, Visualization 7, and Visualization 8). These are pointed by arrows on the global view. The dashed-line indicates the network plane. The red arrow points toward a 12 × 43 × 96 µm3 object and the blue arrow toward a 80 × 80 × 100 µm3 object.

Download Full Size | PDF

To analyse the performances of our setup and 3D reconstruction algorithm to recover unambiguously biological objects, a comparative study with 2D reconstructions from 2D lensfree data acquired at the normal illumination (φ = 0°) was conducted. The algorithm used for performing the 2D reconstructions has been presented in [2], and has already shown, as a 2D lensfree imaging application, its ability to recover single cells and organoids while removing the twin images [2, 5].

Figure 10 shows the 2D reconstruction of a region of the studied field of view. We can observe the efficiency of the reconstruction algorithm to recover the biological scene showing branching networks, single organoids and cells. To perform the comparisons with the 3D reconstruction, three objects were selected from the objects highlighted Fig. 8 and 9: two cells (see ROI 1 and ROI 2 on Fig. 10) and one organoid (see ROI 3 on Fig. 10). Two axial cut profiles respectively in x and y directions were taken for both the 2D and 3D reconstructions, and one cut profile in the z direction for the 3D reconstruction. These cut profiles are illustrated on the three graphs on Fig. 10. As a first observation, we can see that axial profiles are globally equivalent in 2D and 3D, allowing to measure approximatively cell sizes of 15 µm for ROI 2 and 30 µm for ROI 1 considering the width of the central lobe of the profiles (the secondary lobes are known to be twin artefacts), which is consistent with typical cell sizes. For the organoid, the profiles are more erratic but we can distinguish a kind of a plateau curve, the width of which is measured at 90 µm. As expected, the observations are different in the z direction and the resolution is clearly degraded: whereas the objects must have a global isotropic behaviour in terms of sizes in 3D, our measures give 80 µm for the cell in ROI 1, 115 µm for the cell in ROI 2 and 200 µm for the organoid in ROI 3. Moreover, the disparities of the measures for the cells tend to show that the apparent resolution highly depend on the position of the object in the field of view: the more we go away from the centre (from ROI 1 to ROI 2), the worse is the resolution in the z direction. Note that the term of ”resolution” that we employ has not to be confused with the strict definition of optical resolution. These are empirical measures that allow us to conclude that single biological objects are effectively and unambiguously identified in the 3D reconstruction.

Uncertainties in the calibration of the setup and the projection model can explain some arte-facts in the image and the previous quantifications of Fig. 10. For example we did not take into account the stretching of holograms in the registration of the projections which depends of the lighting direction. Thus some misalignment can be present after the registration step. Moreover uncertainties remain on these lighting directions, due to the multiple refractions in between the different propagating media (air, culture medium, sensor glass, etc.) which change the illumination orientations. The induced artefacts then get worse as the distance from the center increases since the mistakes on the objects positioning grow with this distance. That is why the artefactual bow shape around the single objects get worse on the side, as one can see comparing Fig. 8 and 9 and the profiles on Fig. 10.

 figure: Fig. 10

Fig. 10 2D reconstruction from 2D lensfree data acquired on the 3D culture of prostatic cells RWPE1 presented on Fig. 5. Comparisons of two axial cut profiles respectively in x and y directions taken on both 2D and 3D reconstructions, and one cut profile in the z direction taken on the 3D reconstruction, for three biological objects: two cells, one identified on Fig. 8 (ROI 1 in red) and one identified on Fig. 9 (ROI 2 in blue) and one organoid identified on Fig. 9 (ROI 3 in green). The dashed and solid curves correspond respectively to the 2D and 3D reconstruction cut profiles.

Download Full Size | PDF

These critical points highlighted by our demonstrative results will be addressed in future works to allow us a precise quantification of the performances of 3D lensfree imaging, e.g. the effective optical resolution in all directions.

4. Conclusion

In this paper, we have presented a novel technique for the imaging of 3D organoid cultures: lensfree diffractive tomography. Based on lensfree imaging and digital holography principles, we have demonstrated that it is possible to reconstruct large volumes up to several mm3, by multiplying the angles of view around the object of interest. We have developed a reconstruction algorithm based on the direct inversion of the data in the frequency domain using the known Fourier diffraction theorem. We have reconstructed a large volume (4 × 4 × 1.34 mm3) of a culture of RWPE1 prostatic cells in Matrigel. As a result we have been able to observe focused objects of a few tens of microns (cells and network branches).

These first results tend to demonstrate that it is possible to acquire knowledge on the positions and shapes of elements present in 3D cell culture, e.g. spheroids, branching and single cells, on a large and thus relevant volume (several mm3). Acquiring such information is of prior importance for the biologists in the realms of e.g. embryogenesis, angiogenesis and tumorigenesis. In particular the ability to detect isolated single cell distant from the main cell structure, since these cells may present a metastatic phenotype of interest.

References and links

1. N. Scherf and J. Huisken, “The smart and gentle microscope,” Nature Biotechnol. 33(8), 815–818 (2015). [CrossRef]  

2. S. Vinjimore Kesavan, F. Momey, O. Cioni, B. David-Watine, N. Dubrulle, S. Shorte, E. Sulpice, D. Freida, B. Chalmond, J. M. Dinten, X. Gidrol, and C. Allier, “High-throughput monitoring of major cell functions by means of lensfree video microscopy,” Sci. Rep. 4, 5942 (2014). [PubMed]  

3. I. Pushkarsky, Y. Liu, W. Weaver, T.-W. Su, O. Mudanyali, A. Ozcan, and D. Di Carlo, “Automated single-cell motility analysis on a chip using lensfree microscopy,” Sci. Rep. 4, 4717 (2014). [PubMed]  

4. G. Zheng, S. A. Lee, Y. Antebi, M. B. Elowitz, and C. Yang, “The ePetri dish, an on-chip cell imaging platform based on subpixel perspective sweeping microscopy (SPSM),” Proc. Natl. Acad. Sci. U.S.A. 108(18), 16889–16894. (2011). [CrossRef]   [PubMed]  

5. M. E. Dolega, C. Allier, S. V. Kesavan, S. Gerbaud, F. Kermarrec, P. Marcoux, J.-M. Dinten, X. Gidrol, and N. Picollet-D’hahan, “Label-free analysis of prostate acini-like 3D structures by lensfree imaging,” Biosens. Bioelectron. 49, 176–183, (2013). [CrossRef]   [PubMed]  

6. S. O. Isikman, W. Bishara, S. Mavandadi, F. W. Yu, S. Feng, R. Lau, and A. Ozcan, “Lens-free optical tomographic microscope with a large imaging volume on a chip,” Proc. Natl. Acad. Sci. U.S.A. 108(18), 7296–7301. (2011). [CrossRef]   [PubMed]  

7. S. O. Isikman, A. Greenbaum, W. Luo, A. Coskun, and A. Ozcan, “Giga-pixel lensfree holographic microscopy and tomography using color image sensors,” PLOS one 7(9), e45044 (2012). [CrossRef]   [PubMed]  

8. E. Wolf, “Three-Dimensional structure determination of semi-transparent objects from holographic data,” Opt. Commun. 1(14), 153–156 (1969). [CrossRef]  

9. Y. Sung, W. Choi, C. Fang-Yen, K. Badizadegan, R. R. Dasari, and M. S. Feld, “Optical diffraction tomography for high resolution live cell imaging,” Opt. Express 17(11), 266–277 (2009). [CrossRef]   [PubMed]  

10. O. Haeberlé, K. Belkebir, H. Giovaninni, and A. Sentenac, “Tomographic diffractive microscopy: basics, techniques and perspectives,” J. Mod. Opt. 57(9), 686–699 (2010). [CrossRef]  

11. Y. Cotte, F. Toy, P. Jourdain, N. Pavillon, D. Boss, P. Magistretti, P. Marquet, and C. Depeursinge, “Marker-free phase nanoscopy,” Nature Photon. 7, 113–117 (2013). [CrossRef]  

12. A. Kak and M. Slaney, Principles of Computerized Tomographic Imaging (IEEE Press, 1988).

13. J. R. Fienup, “Phase retrieval algorithms: a comparison,” Appl. Opt. 21(15), 2758–2769 (1982). [CrossRef]   [PubMed]  

Supplementary Material (8)

NameDescription
Visualization 1: AVI (4403 KB)      Full reconstructed volume (3D view flyover)
Visualization 2: AVI (2308 KB)      3D view of the full reconstructed volume (3D view flyover)
Visualization 3: AVI (7959 KB)      Crop 1 of the reconstructed volume (3D view flyover)
Visualization 4: AVI (3226 KB)      ROI1 of the crop 1 (3D view flyover)
Visualization 5: AVI (2918 KB)      ROI2 of the crop 1 (3D view flyover)
Visualization 6: AVI (6313 KB)      ROI1 of the crop 2 (3D view flyover)
Visualization 7: AVI (8186 KB)      Crop 2 of the reconstructed volume (3D view flyover)
Visualization 8: AVI (4742 KB)      ROI2 of the crop 2 (3D view flyover)

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (10)

Fig. 1
Fig. 1 Left-hand side - Experimental bench dedicated to lensfree diffractive tomography. Right-hand side - Optical schema of the system. The semi-coherent incident plane wave Uinc passes through the sample volume. Each element of the volume diffracts the incident plane, behaving like secondary spherical sources, creating a diffracted wave Udif. The sensor records the intensity of their summation: Itot = |Utot|2 = |Uinc + Udif|2
Fig. 2
Fig. 2 Schema of three different acquisitions at the illumination angles φ j 1, φ j 2 and φ j 3, with (j1, j2, j3) ∈ ⟦1, N3, where N is the total number of acquisitions.
Fig. 3
Fig. 3 Illustration of the Fourier diffraction theorem with the notations of Fig. 2. The 3D frequency space of the object of interest f is mapped with the 2D Fourier transform of the projections U d i f j 1, U d i f j 2 and U d i f j 3. An example of a Fourier region which is actually filled by the algorithm is given in the medallion in the 3D frequency domain part of the figure with the parameters used in section 3 to reconstruct a given region of interest: a volume of 1.34 × 1.34 × 1.34 mm3 with a voxel size of 1.67 × 1.67 × 3.34 µm3 at λ = 520 nm with an illumination angle varying from −30° (blue) to 30° (green) with a step angle of 5°. The red cap corresponds to the region map with the normal illumination.
Fig. 4
Fig. 4 Illustration of the steps to get an approximation of U d i f j for a given j.
Fig. 5
Fig. 5 3D lensfree data of a culture of prostatic cells RWPE1 in Matrigel. 61 angles of view were acquired from −30° to 30° with an angular pitch of 1° around the x axis. The illumination wavelength is 520 nm. A region of interest (ROI) is extracted on each projection and centered by registration. This aims at performing a localized reconstruction to save computation from RAM overflows. The centering has to be taken into account in the algorithm (Eq. (5)).
Fig. 6
Fig. 6 Extraction of 9 ROIs of size 1.34 × 1.34 mm2 (800 × 800 pixels) for the piecewise 3D reconstruction of the 3D culture of prostatic cells RWPE1 presented on Fig. 5.
Fig. 7
Fig. 7 Piecewise 3D reconstruction (imaginary part of f) of a large volume (4 × 4 × 1.34 mm3) of the culture of prostatic cells RWPE1 from tomographic data presented on Fig. 5. (a) 3 sectional views - xy (center), xz (right) and yz (bottom) - are shown, each one focused at a given distance d = zs − z from the sensor. (b) 3D views of the reconstruction from different observation points (see Visualization 1 and Visualization 2). The dash red curves indicate the tilt of the network relative to the horizontal plane. The red arrows point out isolated objects that are not focused at the same altitude as the network.
Fig. 8
Fig. 8 ROI at the center of the piecewise 3D reconstruction of the volume of Fig. 7. (a) 3 sectional views - xy (center), xz (right) and yz (bottom) - and a 3D view of the ROI. (a) same representation rules on two specific organoids at the full resolution of the reconstruction (see Visualization 3, Visualization 4, and Visualization 5). These are pointed by arrows on the global view. The dashed-line indicates the network plane. The red arrow points toward a 33 × 38 × 42 µm3 object and the blue arrow toward a 16 × 17 × 40 µm3 object.
Fig. 9
Fig. 9 ROI at the to right corner of the piecewise 3D reconstruction of the volume of Fig. 7. (a) 3 sectional views - xy (center), xz (right) and yz (bottom) - and a 3D view of the ROI. (b) same representation rules on two specific organoids at the full resolution of the reconstruction (see Visualization 6, Visualization 7, and Visualization 8). These are pointed by arrows on the global view. The dashed-line indicates the network plane. The red arrow points toward a 12 × 43 × 96 µm3 object and the blue arrow toward a 80 × 80 × 100 µm3 object.
Fig. 10
Fig. 10 2D reconstruction from 2D lensfree data acquired on the 3D culture of prostatic cells RWPE1 presented on Fig. 5. Comparisons of two axial cut profiles respectively in x and y directions taken on both 2D and 3D reconstructions, and one cut profile in the z direction taken on the 3D reconstruction, for three biological objects: two cells, one identified on Fig. 8 (ROI 1 in red) and one identified on Fig. 9 (ROI 2 in blue) and one organoid identified on Fig. 9 (ROI 3 in green). The dashed and solid curves correspond respectively to the 2D and 3D reconstruction cut profiles.

Equations (12)

Equations on this page are rendered with MathJax. Learn more.

U d i f ( r ) = 1 4 π O b j e c t f ( r ) U i n c ( r ) exp ( i k 0 r r ) r r d 3 r = 1 4 π ( f . U i n c ) * h
f ( r ) = k 0 2 ( ( n ( r ) n 0 ) 2 1 )
f ^ ( α j , β j , γ j ) = 4 i π w exp ( 2 i π w z s ) H z s { k 0 j ; λ 0 } U ^ d i f j ( u , v ; z s )
{ p 0 j = 0 q 0 j = sin φ j m 0 j = cos φ j
g ^ ( u ) = g ( x ) exp ( 2 i π u x ) d x
f = 3 D 1 [ S ( H z s j 2 D [ M 0 j U d i f j ] ) | { k 0 j | j = 1 N } ]
M 0 j ( x , y ) = exp ( 2 i π ( n 0 p 0 j λ 0 x + n 0 q 0 j λ 0 y ) )
U d i f j I ¯ t o t j exp ( i k 0 j r ) U i n c j
{ x 0 j = 0 y 0 j = z s tan φ j
f = 3 D 1 [ S ( H z s j D 0 j 2 D [ M 0 j U d i f R O I j ] ) | { k 0 j , x 0 j , y 0 j | j = 1 N j } ]
D 0 j ( u , v ) = exp ( 2 i π ( u x 0 j + v y 0 j ) )
M 0 j ( x , y ) = exp ( 2 i π ( n 0 p 0 j λ 0 ( x + x 0 j ) + n 0 q 0 j λ 0 ( y + y 0 j ) ) ) .
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.