Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Multi-line fluorescence scanning microscope for multi-focal imaging with unlimited field of view

Open Access Open Access

Abstract

Confocal scanning microscopy is the de facto standard modality for fluorescence imaging. Point scanning, however, leads to a limited throughput and makes the technique unsuitable for fast multi-focal scanning over large areas. We propose an architecture for multi-focal fluorescence imaging that is scalable to large area imaging. The design is based on the concept of line scanning with continuous ‘push broom’ scanning. Instead of a line sensor, we use an area sensor that is tilted with respect to the optical axis to acquire image data from multiple depths inside the sample simultaneously. A multi-line illumination where the lines span a plane conjugate to the tilted sensor is created by means of a diffractive optics design, implemented on a spatial light modulator. In particular, we describe a design that uses higher order astigmatism to generate focal lines of substantially constant peak intensity along the lines. The proposed method is suitable for fast 3D image acquisition with unlimited field of view, it requires no moving components except for the sample scanning stage, and provides intrinsic alignment of the simultaneously scanned focal slices. As proof of concept, we have scanned 9 focal slices simultaneously over an area of 36 mm2 at 0.29 µm pixel size in object space. The projected ultimate throughput that can be realized with the proposed architecture is in excess of 100 Mpixel/s.

© 2019 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

In the current era of big data analysis, instrumentation to generate massive amounts of image data is in high demand. For high throughput screening in biology, or for novel computer aided medical diagnoses in the field of digital pathology, there is a need for fluorescence imaging of tissues over large fields of view (~few cm), in 3D (up to hundred layers of µm thickness), and at cellular resolution (~1 µm) [13]. This can be used, for example, in immunofluorescence or fluorescence in situ hybridization (FISH) studies.

The de facto standard modality for fluorescence imaging is scanning confocal microscopy [4], because the optical sectioning capability enables high contrast. The underlying point scanning technique has a limited throughput, and the imaged area is limited by the Field Of View (FOV) of the microscope objective. Parallelization is a strategy to increase throughput, as then the space-bandwidth-time product is increased [5]. For example, in spinning disk microscopy, a large number of points is scanned in parallel [6,7]. Wide field structured illumination has also been proposed as a technique for high throughput imaging [8,9], where the loss in resolution of lower Numerical Aperture (NA) objectives for increasing the FOV, is compensated by the use of structured illumination. Throughput can also be increased by scanning multiple depth layers in parallel using an illumination with multiple foci [10,11]. These techniques require the distribution of the emitted light over several detector arms to apply a pinhole conjugate to the foci that scan the specimen at different depths. The necessary beam splitters used in these approaches result in a loss in collected fluorescence signal strength by a factor equal to the number of scanned layers. Another throughput enhancing technique is the use of a line illumination instead of a spot illumination in combination with a line sensor [1214]. The pinhole for achieving optical sectioning must then be replaced by a slit, which goes at the expense of a small loss in the optical sectioning capability [13,15]. Line scanning can be combined with multi-focus scanning [16], but the proposed method suffers from the same signal losses induced by splitting the beam in the detection path as the point scanning based multi-focus systems.

Scanning large, cm2 sized, areas with all mentioned systems can be accomplished by using ‘mosaic’ or ‘step-and-stitch’ scanning. The most favorable method for scanning such large areas, however, is continuous ‘push broom’ scanning with a line sensor, because of its mechanical simplicity and reduced need for stitching [17,18]. This scanning approach is naturally compatible with confocal line illumination [19]. Stage-scanning instead of beam scanning makes for a system with a minimum number of moving optical components. A line scanning system can be extended by replacing the line sensor with an area sensor. This gives additional freedom for hyper spectral scanning approaches [20,21].

In this paper we propose a multi-focal multi-line scanning fluorescence microscope for efficient 3D imaging over large, cm2 sized, scanning areas. Figure 1 shows the essentials of the scanner concept. The architecture is based on the use of an area image sensor that is tilted with respect to the optical axis, inspired by [22]. The fluorescent specimen is scanned with a set of illumination lines, that are oriented perpendicular to the plane spanned by the optical axis and the scan direction, are focused at different depths inside the specimen, and that are optically conjugate to rows of pixels of the image sensor. This makes it possible to apply confocal slit apertures digitally, post-acquisition. Another advantage is that the line foci used for scanning are laterally separated, avoiding losses in the collected fluorescence signal by beam splitters in existing approaches [10,11].

 figure: Fig. 1.

Fig. 1. Schematic side view of the architecture for a multi-line fluorescence scanning microscope for multi-focal imaging with an unlimited field of view. For clarity we illustrate the concept using only two focal layers. The imaging path is shown in green, the illumination path is shown in blue. A diffractive optical element (DOE) is used for generating a set of parallel scan lines, focused at equidistant planes. Note that in this side view, the lines are orthogonal to the drawing. (a) The tube lens and objective lens form a telecentric optical system such that the tilted sensor has a tilted conjugate plane in object space. The rows on the area sensor conjugate to the line foci are used to capture the image data from the different focal slices scanned by the line foci simultaneously. (b) The sample is scanned in a continuous ‘push broom’ scanning fashion to obtain a multi-focal image with a field of view that is in principle unlimited.

Download Full Size | PDF

The major challenge in realizing this new scanning concept lies in the area of PSF engineering, in which many approaches targeting the imaging light path, often in the context of single-molecule imaging [2328], as well as the illumination light path, in particular light sheet based techniques [29,30], have been proposed. In the current case the single laser beam must be transformed into a set of parallel scan lines projected at different depths inside the sample. Splitting a beam with a diffraction grating into a set of sub-beams with near equal intensity that are focused by the objective lens into a set of equidistant scanning spots in the scan ($y$) direction is an obvious possibility. In this paper we show how to extend the use of diffractive optical elements to first add defocus such that the set of spots scan the sample at equidistant focal layers, and to second modify the spots to lines orthogonal to the plane spanned by the optical ($z$) axis and scan ($y$) axis. Moreover, we introduce a design method for such lines to have uniform intensity along the lines. Secondary challenges in realizing the scanner concept lie in calibration and alignment of the diffractive optical element with the system, in particular with the pixel rows of the image sensor, and with the need to correct for aberrations in the illumination light path. Minimizing the spherical aberration arising from the finite conjugate imaging at different depths is an issue that also must be addressed.

This article has the following structure. In the theory section, the illumination PSF engineering approach to generating the set of equidistant multi-focal scan lines will be described in detail. The experimental methods section starts with a description of our experimental setup. After this, the use of a Spatial Light Modulator (SLM) as diffractive optical element in the setup will be discussed, and, finally, the image metric based aberration correction that is implemented, will be described. In the experimental results section, a quantitative characterization of the realized illumination PSF is presented; images of immunofluorescently stained cells and tissues and of cells stained with Fluorescence in situ Hybridization (FISH) are shown; and a quantitative description of the photo-electron yield of the system, needed to assess the potential optical throughput, is provided. The article is concluded with a discussion of further extension of this scanning platform to higher resolutions and to multi-color imaging.

2. Theory

2.1 Design for generating a set of defocused points

The first step in engineering the desired illumination pattern is to split the illumination laser beam into a set of sub-beams of near equal power, such that the sub-beams form a set of distinct foci after focusing by the lens. Each focus must be a line and the total set of foci must have an equidistant spacing in the scan ($y$) direction and in the focus ($z$) direction. Our design for this is a diffractive structure, often referred to as a ‘diffractive optical element’.

Our design approach is based on the method used earlier by one of us [28]. For the sake of completeness, the essentials of the method are repeated here. The diffractive structure is assumed to be placed in a plane conjugate to the pupil plane and is described as a thin surface, locally altering the phase of the incoming light, independent of the angle of incidence. Then the complex amplitude of the incoming beam is modified by the transmission function $T\left ( \vec {\rho } \right ) = \exp \left ( 2\pi iW(\vec {\rho })/\lambda \right )$, where $\vec {\rho }$ is the normalized pupil coordinate and $W\left ( \vec {\rho } \right )$ is the phase profile added to the incident beam. This phase profile is described by:

$$W\left( \vec{\rho} \right) = f\left( \textrm{mod}\left\lbrack \frac{K\left( \vec{\rho} \right)}{\lambda} \right\rbrack \right),$$
where $\lambda$ is the illumination wavelength, $f(t)$ the so-called profile function with $t \in \lbrack 0,1)$, and $K\left ( \vec {\rho } \right )$ the so-called zone function. It appears that the incident beam is split into diffraction orders with integer index $m$ with amplitude:
$$C_{m} = \int_{0}^{1}{dt\exp\left( - 2\pi imt \right)\exp\left( \frac{2\pi if\left( t \right)}{\lambda} \right)},$$
depending only on the profile function, and phase:
$$W_{m}\left( \vec{\rho} \right) = mK\left( \vec{\rho} \right),$$
depending only on the zone function. For a diffraction grating, the zone function is a linear function of the pupil position, leading to an added aberration that only consists of wavefront tilt. This is sufficient for splitting the incoming beam into a set of diffraction orders (the required set of sub-beams), that will result in a set of foci with an equidistant spacing in the lateral (scan, $y$) direction. In our case, however, it is desired to have an equidistant spacing of the diffraction orders in the axial (focus, $z$) direction as well. We therefore choose a zone function that is a sum of tilt and defocus:
$$K\left( \vec{\rho} \right) = \mathrm{\Delta}y\frac{\textrm{NA}{}\rho_{y}}{\lambda} + \mathrm{\Delta}z\frac{\sqrt{n^{2} - \textrm{NA}{}^{2} |\vec{\rho}|^{2}}}{\lambda},$$
with $\mathrm {\Delta }y$ the lateral spacing of the foci, $\mathrm {\Delta }z$ the axial spacing of the foci, $n$ the refractive index of the sample medium, and NA the microscope objective Numerical Aperture.

The next step in the design is to find a profile function $f(t)$ that gives a power distribution over the diffraction orders $\left | C_{m} \right |^{2}$ that is as uniform as possible. This problem is in general related to the problem of finding a phase function that leads to a desired intensity distribution, and can for example be solved using the Gerchberg-Saxton algorithm [31]. We use a parameterization of the profile function in terms of a finite Fourier series:

$$f\left( t \right) = \sum_{n = 1}^{N}{a_{n}\cos\left( 2\pi nt \right)},$$
which advantageously results in a smooth and band-limited profile function, and is characterized by the $N$ free parameters $a_{n}$. The cost function is defined as
$$Z\left( a_{0},\ldots,\;a_{N} \right) = \sum_{m ={-} M}^{M}\left( \left| C_{m} \right|^{2} - \eta_{m} \right)^{2},$$
where an integer $M \geq N$ is used, and where $\eta _{m}$ represents the desired set of diffraction efficiencies. An even distribution of power over the first $2K + 1$ diffraction orders is obtained when $\eta _{m}=1/\left ( 2K + 1 \right )$ for $m \leq K$ and zero otherwise. A generic search algorithm can be used for the minimization problem, such as implemented in the function fminsearch in MATLAB. In our experiments we have implemented a solution for $K = 4,N = 5,M = 10$, which results in a profile function that gives 97 % of the power in the 9 diffraction orders, where the power in the individual orders varies by less than 3 %, see Fig. 2.

 figure: Fig. 2.

Fig. 2. Demonstration of grating with tilt and defocus. (a) The zone function with $\Delta y = 10\lambda$ and $\Delta{z} = 10\lambda$. These numbers are chosen for demonstration purpose, and do not represent a realistic scenario. (b) The profile function optimized for 9 lines. (c) The resulting distribution of power over the diffraction orders. (d) The corresponding aberration function.

Download Full Size | PDF

2.2 Design for lines with uniform intensity

So far, the diffractive structure is capable of producing a set of point foci, equidistantly spaced in the scan ($y$) direction as well as in the focus ($z$) direction. The next step is to modify the set of point foci into a set of line foci, where the direction of the lines is orthogonal to the $yz$-plane of the original point foci. This can be accomplished by adding astigmatism to the aberration profile, giving a total aberration profile:

$$W\left( \vec{\rho} \right) = W_{yz}\left( \rho_{x},\rho_{y} \right) + W_{x}\left( \rho_{x} \right),$$
where $W_{yz}\left ( \rho _{x},\rho _{y} \right )$ is the aberration profile of the grating structure, according to Eqs. (1), (4), and (5), and where $W_{x}\left ( \rho _{x} \right )$ represents astigmatism. This aberration function depends only on the normalized pupil coordinate $\rho _{x}$ as the intended line foci are oriented in the lateral plane orthogonal to the scan (y) direction. The astigmatic aberration can be generated by a refractive optical element such as a cylindrical lens or by the same diffractive optical element that produces the grating structure.

It appears that a simple parabolic astigmatic aberration profile $W_{x}\left ( \rho _{x} \right ) = \frac {1}{2}a{\rho _{x}}^{2}$ with $a$ a coefficient, results in a line focus where the peak intensity and line width vary along the line. Here, we add higher order astigmatism in order to overcome these issues, generating focal lines with substantially constant peak intensity and line width along the line. The optical design for these improved astigmatic focal lines starts with a simplified 1D diffraction model for computing the impact of a general aberration profile $W_{x}\left ( \rho _{x} \right )$ on the focal line shape. Consider an infinitesimal strip in the back focal plane at location $\rho _{x}$ with width $R d\rho _{x}$ and height $2R\sqrt {1 - {\rho _{x}}^{2}}$, with $R$ the pupil radius as indicated in Fig. 3(a). This strip generates a plane wave that makes an angle $\alpha$ with the optical axis as indicated in Fig. 3(b), given by:

$$\sin\alpha = \frac{1}{R}\frac{dW_{x}}{d\rho_{x}}.$$
This plane wave is focused to a point in the field of view a distance $x$ from the optical axis given by:
$$x = F\sin\alpha = \frac{1}{\textrm{NA}{}}\frac{dW_{x}}{d\rho_{x}},$$
with $F$ the focal length of the objective lens, which we assume to be aplanatic and where we have used that $\textrm {NA}{} = R/F$ . The strip width in the front focal plane as illustrated in Fig. 3(c) is given by:
$$dx = \frac{d\left( F\sin\alpha \right)}{d\rho_{x}}d\rho_{x} = \frac{1}{\textrm{NA}{}}\frac{d^{2}W_{x}}{d{\rho_{x}}^{2}}d\rho_{x}.$$
Equation (9) provides a unique mapping of pupil coordinate $\rho _{x}$ to field position $x$ if the second order derivative of $W_{x}$ as a function of $\rho _{x}$ does not change sign. Now, the intensity in the front focal plane according to the 1D diffraction model is:
$$I_{f}(x,\;y) = \left( \frac{1}{\textrm{NA}{}}\frac{d^{2}W_{x}}{d{\rho_{x}}^{2}} \right)^{- 1} \cdot \left| U_{f}(x,\;y) \right|^{2},$$
where the first factor is the ratio between the width of the strip in the back focal plane and the width of the strip in the front focal plane. The second factor is the intensity resulting from diffraction in the $\rho _{y}$-direction:
$$U_{f}(x,\;y) = \frac{\textrm{NA}{}}{\lambda}\int_{- \sqrt{1 - {\rho_{x}}^{2}}}^{\sqrt{1 - {\rho_{x}}^{2}}}{d\rho_{y} A\left( \rho_{x},\rho_{y} \right)\exp\left( - \frac{2\pi i\rho_{y}y \textrm{NA}{}}{\lambda} \right)},$$
where the mapping from pupil coordinate $\rho _{x}$ to field position $x$ is implicitly taken into account, and where $A(\rho _{x},\rho _{y})$ is the amplitude profile of the beam incident on the diffractive structure. It is mentioned that the 1D diffraction model can also be derived from the standard 2D diffraction integral over the pupil plane if we apply the stationary phase approximation to the integral over the pupil coordinate $\rho _{x}$.

 figure: Fig. 3.

Fig. 3. Illustration of the proposed geometry for the aspherical cylindrical lens design. (a) In the back focal plane, we have an circular aperture which is illuminated by a beam profile $A\left ( \rho _{x},\rho _{y} \right )$. We introduce an astigmatic aberration function $W$ such that every strip in the back focal plane there is a corresponding strip in the front focal plane (red boxes in all subfigures). (b) Side view of the optical system, with the back focal plane (bfp), the objective lens (obj) and the front focal plane (ffp). (c) The strips in the front focal plane have a non-constant width $dx$. By changing the shape of the $W$, $dx$ can be modified to obtain a uniform distribution of power.

Download Full Size | PDF

Assuming that the incident amplitude profile is flat, $A\left ( \rho _{x},\rho _{y} \right ) = A,$ we find that:

$$U_{f}\left( x,\;y \right) = \frac{2A \textrm{NA}{}}{\lambda}\sqrt{1 - {\rho_{x}}^{2}}\textrm{sinc}\left( 2\pi\sqrt{1 - {\rho_{x}}^{2}}\frac{y\textrm{NA}{}}{\lambda} \right),$$
with $\textrm {sinc}\left ( t \right ) = \sin \left ( t \right )/t$. The focal line intensity according to Eq. (11) is:
$$I_{f}\left( x,\;y \right) = \left( \frac{1}{\textrm{NA}{}}\frac{d^{2}W_{x}}{d{\rho_{x}}^{2}} \right)^{- 1}\frac{4A^{2} \textrm{NA}{}^{2}}{\lambda^{2}}\left( 1 - \rho_{x}^{2} \right)\left[\textrm{sinc}\left( 2\pi\sqrt{1 - {\rho_{x}}^{2}}\frac{y\textrm{NA}{}}{\lambda} \right)\right]^{2}.$$
For the simple parabolic phase profile $W_{x}\left ( \rho _{x} \right ) = \frac {1}{2}a{\rho _{x}}^{2}$, corresponding to a cylindrical lens in the paraxial approximation, the pupil coordinate $\rho _{x}$ and field position $x$ are related by $\rho _{x} = \textrm {NA}{} x/a$, and we find a focal line intensity profile:
$$I_{f}\left( x,\;y \right) = \frac{4A^{2}\textrm{NA}{}^{3}}{a\lambda^{2}}\left( 1 - \left[ \frac{\textrm{NA}{} x}{a} \right]^{2} \right){\textrm{sinc}\left( 2\pi\sqrt{1 - \left[ \frac{\textrm{NA}{} x}{a} \right]^{2}}\frac{ y\textrm{NA}{}}{\lambda} \right)}^{2}.$$
The peak intensity decays as $1 - \left ( \textrm {NA}{}x/a \right )^{2}$ along the line, and the line width increases as $1/\sqrt {1 - \left ( \textrm {NA}{}x/a \right )^{2}}$ along the line.

A peak intensity along the line that is independent of $x$ can be achieved if:

$$\frac{d^{2}W_{x}}{d{\rho_{x}}^{2}} \propto 1 - {\rho_{x}}^{2}.$$
This results in an aberration profile $W_{x}\left ( \rho _{x} \right )$ that is of fourth order in $\rho _{x}$. Ignoring constant phase offsets, and requiring a symmetric line structure centered at the optical axis (no odd terms in $\rho _{x}$), we arrive at:
$$W_{x}\left( \rho_{x} \right) = \frac{3}{4}w\textrm{NA}{}\left\lbrack {\rho_{x}}^{2} - \frac{1}{6}{\rho_{x}}^{4} \right\rbrack,$$
with $l$ an integration constant. The relation between pupil coordinate $\rho _{x}$ and field position $x$ according to Eq. (9) is:
$$x = \frac{3}{2}l\left\lbrack \rho_{x} - \frac{1}{3}{\rho_{x}}^{3} \right\rbrack.$$
At the pupil rim $\rho _{x} = 1$ we find that $x = l$, implying that the total line length is $2l$. Equation (18) may be inverted to:
$$\rho_{x} = 2\cos\left( \frac{\pi}{3} + \frac{1}{3}\arccos\left( \frac{x}{l} \right) \right).$$
It is mentioned that, although the peak intensity is constant along the line, the line width still depends somewhat on field position $x$. An alternative requirement that might be pursued is the requirement that the total power is distributed equally along the line, i.e. $\int dyI_{f}(x,\;y)$ is constant. It then follows that:
$$\frac{d^{2}W_{x}}{d{\rho_{x}}^{2}} \propto \int dy|U_{f}(x,\;y)|^{2} \propto \sqrt{1 - {\rho_{x}}^{2}},$$
which can be solved in a similar way as before, leading to:
$$W_{x}\left( \rho_{x} \right) = w\frac{2\textrm{NA}{}}{\pi}\left\lbrack \rho_{x}\arcsin(\rho_{x}) - \frac{1}{3}\left( 1 - {\rho_{x}}^{2} \right)^{\frac{3}{2}} + \sqrt{1 - {\rho_{x}}^{2}} \right\rbrack$$
In our experiments we have opted for keeping the peak intensity along the line as constant as possible, which is optimal in the limit of a small confocal detection slit.

We have tested the theoretical analysis with our 1D diffraction model by computing the shapes of line foci computed numerically with standard 2D diffraction by a circular aperture. Figure 4 shows an evaluation of the obtained results for a design line length $2l = 500{\lambda /\textrm {NA}{}}$. Figure 4(a) shows the result for an aberration function that consists of lowest order astigmatism only, as e.g. created by the use of a cylindrical lens. The peak intensity over the line decreases to zero and the line broadens towards the edges. The full width at half maximum of this line is $1.42l$, i.e. just 71 % of the total intended line length $2l$. The practically usable part of the line is even less than this as a drop of 50 % in peak intensity is already quite high. Figure 4(b) shows the result for the aberration function including higher order astigmatism, optimized for uniform peak intensity according to Eq. (17). The peak intensity turns out to be uniform over the full line, as predicted by the 1D diffraction model, except for a small amount of ringing close to the line edges as can be seen in Fig. 4(c). This ringing originates from diffraction at the aperture edges in the $\rho _{x}$-direction, which is not accounted for in the 1D diffraction model. At $x = \pm 220\lambda /\textrm {NA}{}$ ringing is no longer present and the line width is within one Airy unit, as shown in Fig. 4(d). We conclude that about 85 % of the design line length can be used, without a substantial decrease in peak intensity or a substantial increase in line width.

 figure: Fig. 4.

Fig. 4. Calculated intensities in the focal plane using 2D scalar diffraction. A value $l = 256 \lambda /\textrm {NA}{}$ was used. The aperture is assumed to be circular. The red lines indicate one airy unit. (a) A line created using a parabolic aberration function. (b) A line created using an aberration function with higher order astigmatism optimized for uniformity conform Eq. (17). The yellow boxes correspond to the detailed images given in c-e. (c) Close to the edge of the line profile ringing is present. (d) This ringing has disappeared around $x = - 220 \lambda /\textrm {NA}{}$. Here the line width is within one Airy unit. (e) In the center of the line pattern, the line has the smallest line width. The peak is constant over the line.

Download Full Size | PDF

2.3 Image formation theory

Several aspects of the envisioned imaging system have impact on the formal image formation model. First of all, the illumination is with a set of line foci, making the illumination intensity as a function of position in object space:

$$H_{ill}\left( x,\;y,\;z \right) = \sum_{m}^{}{H_{ex}\left( y - m\mathrm{\Delta}y,\;z - m\mathrm{\Delta}z \right)},$$
where $H_{ex}\left ( x,\;z \right )$ is the intensity profile of the line foci, $\mathrm {\Delta }y$ is the lateral separation of the line foci, $\mathrm {\Delta }z$ is the axial separation of the line foci, and the sum is over all participating diffraction orders. Second, the image data is captured at a tilted image sensor, which mixes the lateral and axial coordinates. For a fluorescent object $F\left ( x,\;y,\;z \right )$ recorded for a scan coordinate $y_{s}$, the intensity recorded at the image sensor is:
$$I\left( Mx',My',\;y_{s} \right) = \int_{}^{}{dxdydz\: H_{em}\left( x' - x,\;y' - y, - z + \frac{\mathrm{\Delta}z}{\mathrm{\Delta}y}y' \right)F\left( x,\;y + y_{s},\;z \right)H_{ill}\left( x,\;y,\;z \right)},$$
where $H_{em}\left ( x,\;y,\;z \right )$ is the emission Point Spread Function (PSF), and $M$ is the lateral magnification. This may be written as a sum over line images:
$$I\left( Mx',My',\;y_{s} \right) = \sum_{m}^{}{I_{m}\left( x',\;y' - m\mathrm{\Delta}y,\;y_{s} \right)},$$
with (substitute $y''=y'-m\mathrm {\Delta }y$):
$$\begin{aligned}I_{m}\left( x',\;y'',\;y_{s} \right) &= \int_{}^{}{dxdydz\: H_{em}\left( x' - x ,\;y'' - y+ y_{s}, - z + \frac{\mathrm{\Delta}z}{\mathrm{\Delta}y}y'' \right)H_{ex}\left( y - y_{s},\;z \right)}\\ &\times F\left( x,\;y + m\mathrm{\Delta}y,\;z + m\mathrm{\Delta}z \right), \end{aligned}$$
Taking into account that only a small range of $y''$ values around the central line position at $y''=0$ give rise to substantially non-zero intensities we may approximate this as:
$$\begin{aligned}I_{m}\left( x',\;y'',\;y_{s} \right) &= \int_{}^{}{dxdydz\: H_{em}\left( x' - x,\;y'' - y + y_{s}, - z \right)H_{ex}\left( y - y_{s},\;z \right)}\\ &\times F\left( x,\;y + m\mathrm{\Delta}y,\;z + m\mathrm{\Delta}z \right). \end{aligned}$$
The procedure of digital spatial filtering following a slit shape, possibly, including the collection of extra signal along the scan with a TDI operation can be incorporated by setting $y_{s} = y_{i} - y''$ and integrating over $y''$, where the integration range is $K$ rows of pixels (we take $K = 4$ or $K = 1$). Here $y_{i}$ is the image coordinate in the scan direction. The image coordinate in the field direction is simply found by $x' = x_{i}$. Then we find for the image data at layer $m$:
$$\begin{aligned}J_{m}\left( x_{i},\;y_{i} \right) &= \int_{}^{}{dy''\: I_{m}\left( x_{i},\;y_{i},\;y_{i} + y'' \right)}\\ &= \int_{}^{}{dxdydz\: H\left( x_{i} - x,\;y_{i} - y, - z \right)F\left( x,\;y + m\mathrm{\Delta}y,\;z + m\mathrm{\Delta}z \right)}, \end{aligned}$$
where the total PSF:
$$H\left( x,\;y,\;z \right) = H_{em}\left( x,\;y,\;z \right)\left\lbrack \int_{}^{}{dy''\:H_{ex}\left( - y + y'', - z \right)} \right\rbrack$$
is seen to be the product of the emission PSF with the average of the line excitation PSF averaged over a lateral distance corresponding to $K$ rows of pixels. Close to focus, this integration range is sufficient to capture the entire excitation signal, maximizing signal, and reducing the overall PSF to the emission PSF of the system. This stands in contrast to conventional confocal point scanning microscopes, where the imaging in focus is determined by the excitation PSF. Further away from focus, the integration range is substantially less than the width of the defocused lines. Consequently, the captured signal is reduced and optical sectioning is achieved. The complete sectioning behavior is described by the background function:
$$B\left( z \right) = \int_{}^{}dxdy\: H\left( x,\;y,\;z \right),$$
which gives the contribution to the background of a layer a distance $z$ away from the imaged plane. For non-sectioning widefield imaging $B\left ( z \right ) = 1$, for sectioning modalities $B\left ( z \right )$ is peaked around $z=0$, decaying to zero as $\left | z \right |\rightarrow \infty$.

The illumination lines are characterized experimentally by imaging thin uniform fluorescent layers, for which $F\left ( x,\;y,\;z \right ) \approx \delta \left ( z -z' \right )$, with $\delta \left ( z \right )$ the delta-function, in a static configuration without any scanning ($y_{s} = 0$) but with a tunable defocus $z'$. Then the expected line image is a function of the image coordinate in the scan direction $y'$ and of the defocus $z'$ only:

$$I_{m}\left( x',\;y',\;z' \right) = \int_{}^{}{dxdy H_{em}\left( x' - x,\;y' - y,\;m\mathrm{\Delta}z-z' \right)H_{ex}\left( y, z'- m\mathrm{\Delta}z \right)},$$
which leads to an image that is the average of the emission PSF in the field ($x$) direction, and the convolution of the line excitation PSF and the emission PSF in the scan ($y$) direction. This image function depends is dominated by the broadest of the two functions involved in the convolution, either the line excitation profile or the average of the emission PSF in the field direction. Integrating the obtained images over the lateral coordinate provides a measurement of the background function $B\left ( z \right )$ needed to assess the degree of confocality.

The illumination lines are further characterized experimentally by imaging thick fluorescent layers for which $F\left ( x,\;y,\;z \right )$ is substantially constant, in a static configuration without any scanning ($y_{s} = 0$), and with an additional defocus $z'$ applied to the excitation PSF. In that case the expected line image as a function of $y'$ and $z'$ is given by:

$$T_{m}\left( y',\;z' \right) = \int_{}^{}{dxdydz H_{em}\left( x' - x, - y,\;z' - z \right)H_{ex}\left( y,\;z \right)},$$
where a change of integration variable is used for the integration over the axial coordinate. This through-focus line PSF like function is thus the 2D convolution of the through-focus line excitation PSF and the average of the through-focus emission PSF over the field coordinate. The measured through-focus line shape is determined by the broadest of these two functions.

3. Experimental methods

3.1 Experimental setup

The experimental setup is illustrated in Fig. 5. A Nikon $20 \times$ NA 0.75 Plan Apochromat VC objective lens with a focal length of $F_{ob} = 10$ mm and a custom Nikon tube lens (effective focal length of $F_{tube}=222.4\pm 2.2$ mm) are used, giving a lateral magnification of the imaging light path $M = 22.24$. The axial magnification of the imaging path is $M_{ax}=\chi M^{2}/n = 352$, where a sample refractive index $n=1.5$ is assumed and $\chi = 1.07$ is a non-paraxial correction factor [32,33]. The specimen (slide) is scanned using two stages: a PI M-505 low profile translation stage for positioning of the slide in the field direction and a Newport XM1000 ultra precision linear motor stage for the continuous scanning motion of the sample. For focusing two stages are used to axially translate the objective: a PI M-111 compact micro-translation stage for coarse positioning, and a PI P-721.CL0 piezo nanopositioner for fine positioning.

 figure: Fig. 5.

Fig. 5. Schematic illustration of the multi-focal line confocal scanner prototype. The polarization of the laser beam is aligned to the SLM using a half wave plate (HWP) and filtered using a polarizer (POL). The beam is expanded and projected onto the SLM. The diffracted light is coupled into the objective lens using a filter cube consisting of an excitation filter (EX), emission filter (EM) and a dichroic mirror (DIC) and imaged in the back focal plane of the objective. The fluorescence light is imaged onto the CMOS sensor using the tube lens. Focal lengths $F$ are given in mm. (b). The measured impact on the photon detection efficiency of tilting the camera with respect to the optical axis. For a tilt of 20° the detection efficiency is reduced with 53 % with respect to orthogonal incidence.

Download Full Size | PDF

A Hamamatsu Orca Flash 4.0 v2 camera is used with a pixel pitch of 6.5 µm corresponding to $p=0.29$ µm in object space, which somewhat undersamples the imaging PSF (Nyquist sampling is at $\lambda /4\textrm {NA}{} = 0.20$ µm for $\lambda = 590$ nm). The gain of the camera is measured to be 0.39 e/ADU. The camera is tilted over an angle of $\beta =20$° with respect to the optical axis. The image sensor size in the direction perpendicular to the line sensors is 2048$\times$6.5 µm= 13.3 mm, such that the axial range in image space is $d = 4.4$ mm and the axial range in object space is $d/M_{ax} = 12.5$ µm. The Orca Flash camera has a micro-lens array for higher photon detection efficiency, which makes the sensor less efficient for light that is incident at an angle. The measured loss in light collection efficiency is shown in Fig. 5(b). It appears that tilting the detector 20° reduces the light collection efficiency to 53 % of its nominal value. The camera supports exposure times down to 1 ms. The frame rate, however, is limited by the time required for the rolling shutter to read out the sensor area, leading to a maximum frame rate of 100 fps for a full frame acquisition.

We found the camera to skip some frames when operated at high frame rates, which was solved by reducing the frame rate to 21 fps.

The light source is an Omicron LightHUB-4 equipped with a PhoXX+ 488-100 laser with 100 mW output power and 488 nm wavelength. A polarization maintaining broadband fiber with an achromatic collimator is used to obtain a collimated beam in free space. A beam expander is built using a Thorlabs AC254-50-A, and AC254-200-A achromats for a $4 \times$ magnification. A Liquid Crystal on Silicon (LCoS) Spatial Light Modulator (SLM, Boulder Nonlinear Systems XY Phase 512) is used for the required illumination PSF engineering. The SLM has $N^{2} = 512 \times 512$ pixels and a square aperture of 7.68 mm, resulting in a pixel pitch of 15 µm. A Thorlabs AHWP10M-600 achromatic half wave plate is used to align the polarization of the beam to the SLM and a Thorlabs LPVISE100-A linear polarizer is used to suppress the orthogonal polarization. A relay path from SLM to the back focal plane of the objective lens is made using a Thorlabs AC254-150-A achromat ($F =150$ mm), and an AC254-100-A achromat ($F = 100$ mm giving a magnification of $M = 0.67$). A Semrock LF405/488/532/635-B-000 quad-band dichroic mirror is used to couple in the illumination light into the objective and direct the emission light towards the camera. A National Instruments NI PXIe-6363 data acquisition card (DAQ) is used for synchronization of the hardware. The DAQ gives a start trigger to the camera for each frame, and it provides a high signal to the laser for digital modulation, such that the laser is only switched on during the camera integration period. The acquisition card is synchronized to the output trigger of the scanning stage.

3.2 Fluorescent test samples

Two test slides are used to calibrate and test the scanner setup. First, we use a 1.4 mm thick green plexiglas slide of Chroma that shows strong autofluorescence when excited with blue light. A glass coverslip was attached onto the slide for proper spherical aberration matching. This removes the need for imaging deep into the material, and so prevents scattering and reduces out-of-focus fluorescence. This slide has high fluorescence, is relatively homogeneous and has low bleaching. Second, we used a slide with a spin coated fluorescent layer of Rhodamine on a coverslip. This slide has the advantage of being very thin ($\ll$1 µm), but is relatively inhomogeneous and suffers from bleaching.

Two test samples are used to demonstrate the scanner system. The first slide contains a 5 µm thick human prostate tissue micro array (TMA) section labeled using the Kreatech ERBB2 (17q12) / SE 17 FISH probe (product number KBI-10701). This labeling is used for the detection of amplification of the ERBB2 (also known as HER-2/neu) gene via Fluorescence in situ Hybridization (FISH). As a reference, also the chromosome 17 centromere is labeled using a Satellite Enumeration (SE) probe. The ERBB2 specific FISH probe is direct-labeled with PlatinumBright550. The SE 17 specific FISH probe gene region is direct-labeled with PlatinumBright495. The SE 17 probe is imaged with our setup. The average wavelength of the fluorophore at the detector (taking into account the emission spectrum and the dichroic filter) is 563 nm. The second slide is a stage 3 human rectum cancer sample with immunofluorescence staining. Three different antibodies label Desmin, CD31, and D2-40. In this work, the Desmin protein (labeled with IgG1 M Alexa Fluor 488) is imaged, which is highly expressed in muscle cells. The average wavelength of the fluorophore at the detector (taking into account the emission spectrum and the dichroic filter) is 539 nm.

3.3 Data acquisition and processing

Image data is acquired in a ‘push broom’ scanning fashion [17,18]. The speed of the scan movement is $v = p\cos {\beta /t_{l}}$, where $\beta$ is the sensor tilt angle, $t_{l}$ is the line period and $p$ is the pixel pitch in object (sample) space. Every line period the camera captures a full frame which is transferred to the host computer. A number of regions of interest (ROIs) corresponding to the number of illumination lines are selected, the rest of the image data is deleted. The ROIs have an equal size, span the width of the sensor, and have an equal spacing $s_{r} = Ms_{l}/\cos \beta$ where $s_{l}$ is the tangential illumination line spacing in object space. The selected ROIs consist of either one or four rows of pixels. This functions as a confocal slit detector of width $p\cos \beta$ or $4p\cos \beta$. The data is added in a Time Delayed Integration (TDI) fashion for the four-line acquisition mode, in order to increase signal-to-noise ratio without sacrificing resolution. [34]. The data obtained in this way for the different ROIs correspond to the layers of the simultaneously acquired focal stack. Storage of the image data to the computer memory of layer $m$ is delayed with a time $m = s_{r}t_{l}m$ in order to guarantee lateral alignment of the layers of the focal stack. The data is saved to hard disk after completion of a lane scan.

The following steps are taken in post-processing. First, a fine alignment between the layers is done to compensate for residual misalignments by optimizing the product of all layers on a $1024 \times 1024$ pixels patch with distinct point sources. Second, the image is resampled with a factor $\cos \beta$ in the scan direction to have an image with square pixels. Third, a constant value is subtracted to correct for detector offset. Finally, flat fielding is applied to compensate for residual intensity variations (down to 50 % at the edge of the region of interest) mainly arising from the Gaussian illumination profile of the back aperture of the objective.

A maximum intensity projection along the $z$ direction is used for a 2D visualization of the 3D dataset. For improved dynamic range, a gamma correction is applied. The gray scale image is mapped to RGB by multiplying the gray level with an RGB triplet based on the calculated filtered fluorophore emission spectrum: (0.55, 1, 0.08) for Platinum495 and (0.28, 1.0, 0,15) for Alexa488.

3.4 Spatial light modulator

The so-called space-bandwidth product of the illumination light path (product of illuminated FOV and spatial frequency bandwidth of the illumination pattern) is limited by the number of pixels of the SLM $N.$ Suppose the SLM is imaged by the relay lenses to an aperture of width $w$ in the back focal plane of the objective lens. This provides an illumination NA given by $\textrm {NA}{}_{ill} = w/{2F}_{0}$ with $F_{0}$ the focal length of the objective. The maximum spatial frequency that can be generated by the SLM is $N/2w$, which gives rise to field angles $\eta$ up to $\sin \eta = \lambda N/2w$, resulting in a field of view:

$$\textrm{FOV}{}_{ill} = {2F}_{0}\sin\eta = \frac{N}{2}\frac{\lambda}{\textrm{NA}{}_{ill}}.$$
In other words, the product between $\textrm {FOV}{}_{ill}$ and spatial frequency bandwidth $2\textrm {NA}{}/\lambda$ is fixed by the number of SLM pixels $N$. In the resulting trade-off between $\textrm {FOV}{}_{ill}$ and $\textrm {NA}{}_{ill}$, we have chosen for $\textrm {NA}{}_{ill} = 0.26$ and $\textrm {FOV}{}_{ill} =$ 488 µm.

An illumination line pitch of $s_l=54.5$ µm is used, corresponding to 200 pixel rows on the sensor. The 9 lines span a total distance $8s_l=436$ µm, which conveniently fits in the $\textrm {FOV}_{ill}$. Intentionally some room was left at the edges of the FOV because this area is most sensitive for ghost lines. These lines correspond to spatial frequency content of the aberration profiles higher than the bandwidth defined by the sampling density of the pixel grid of the SLM. These spatial frequencies are folded back into the central field of view.

Using a multi-line imaging approach can potentially lead to optical cross-talk, i.e. fluorescent signal excited by a line beamlet is detected at the pixel rows that are conjugate to the neighboring line beamlet. The sensitivity to cross-talk can be estimated using a geometrical optics based argument. Suppose a line of interest has its focus at the origin $(y,\;z)=(0, 0)$. Light is collected from within a cone with a half angle $\alpha _{det}=\arcsin (\textrm {NA}_{det}/n)$, where $\textrm {NA}_{det}$ is the detection NA and $n$ the sample refractive index. The neighboring line is illuminated with a beam focused at $(y,\;z)=(s_l, -M \tan \beta s_l / M_{ax})$, where $\beta$ is the sensor tilt, and $M$ and $M_{ax}$ the lateral and axial magnification of the imaging path. This beam has a half angle of $\alpha _{ill}=\arcsin (\textrm {NA}_{ill}/n)$, where $\textrm {NA}_{ill}$ is the illumination NA. The $z$ position at which a fluorophore excited by the neighboring illumination beam first enters the cone of the detection NA is given by:

$$z_c = s_l \frac{1 - \tan\alpha_{ill} \tan\beta}{\tan\alpha_{ill} + \tan\alpha_{det}}.$$
Filling in the relevant parameters results in $z_c=72$ µm. Light from such a distance away from focus will in practice be completely suppressed by the confocal detection.

An SLM has a limited diffraction efficiency, which leaves a part of the incoming beam unmodulated, resulting in a focused ‘zero-order’ spot in the sample. The diffraction efficiency is affected by several factors [35]. First, the SLM only works for one polarization direction. A half wave plate and a polarizer were included in the beam expander for polarization alignment and suppression of residual orthogonally polarized light. Additionally, the top of the glass surface will contribute to a plane wave reflection, even though it is coated with an anti-reflection layer. Finally, the limited fill factor of the SLM pixels leave parts of the aperture unmodulated. The line illumination pattern is aligned such that the zero-order spot is centered between two lines to enable spatial filtering at the image sensor, and will therefore have no impact on the final 3D image.

The SLM response curve is described by the phase change $\phi (\epsilon )$ as a function of the digital input gray level $\epsilon \in [0,1)$. A lookup table is required to find the gray level given the desired phase shift. A single lookup table is used across the field of the SLM, implying that non-uniformity of the response function is not taken into account. Coarse pixel-to-pixel variations result in low order aberrations that are corrected later on, fine pixel-to-pixel variations result in a small scatter background that has little impact on the final image.

An image based method is developed to obtain the response curve of the SLM. A binary grating with a period much larger than a single SLM pixel and gray levels $\epsilon _{1}$ and $\epsilon _{2}$ is created, for a range of values $\left ( \epsilon _{1},\epsilon _{2} \right ).$ The power in the diffracted orders is obtained by imaging a diffraction pattern that is projected onto the uniformly fluorescent calibration slide. The space $\left ( \epsilon _{1},\epsilon _{2} \right )$ is sampled randomly instead of linearly to prevent biases arising from bleaching during data acquisition. Then, the ratio between the power in the first and second diffraction order is measured and fitted with the theoretically expected value:

$$R(\epsilon_{1},\epsilon_{2}) = \frac{\sin\left( \frac{1}{2}\left\lbrack \phi\left( \epsilon_{1} \right) - \phi\left( \epsilon_{2} \right) \right\rbrack \right)}{\cos{\left( \frac{1}{2}\left\lbrack \phi\left( \epsilon_{1} \right) - \phi\left( \epsilon_{2} \right) \right\rbrack \right) + C}},$$
where an extra degree of freedom $C$ is introduced to account for the above mentioned zero order spot and possible background and detector offset. The response function $\phi \left ( \epsilon \right )$ is parameterized by a polynomial series of order $L$, and MATLABs fminsearch algorithm is used to fit the polynome coefficients. Figure 6(a) shows the measured values, Fig. 6(b) shows the fit model for the obtained fit parameters, and Fig. 6(c) shows the corresponding SLM response curve $\phi \left ( \epsilon \right )$. We found $L = 3$ to provide enough degrees of freedom to fit the data. The total phase modulation of the SLM exceeds $2\pi$ for the used wavelength, which leaves the freedom to choose a subset of $\epsilon$ values to create the lookup table. We choose the range between $\epsilon = 0.31$ and $\epsilon = 0.90$, as indicated with the red boxes.

 figure: Fig. 6.

Fig. 6. Results of the image based calibration of the SLM response curve. (a) Measured ratio between the first and zeroth diffraction order as a function of the gray levels $\epsilon _{1}$ and $\epsilon _{2}$ of the binary grating structure. (b) Modeled diffraction order ratio for the found fit parameters. (c) The induced phase change as a function of the digital gray level. The red boxes indicate the gray values that are used to create the lookup table and correspond to a $2\pi$ phase change.

Download Full Size | PDF

3.5 Alignment of the spatial light modulator

A careful alignment of the illumination PSF, consisting of the set of multi-focal scan lines, with the pixel grid of the image sensor is required for implementing digital slit pinholes at the imaged line positions in the sensor plane. The reference coordinate system to align the scan lines is given by the optical axis of the objective lens ($z$-axis) and the two principal directions of the pixel grid of the sensor, which correspond to the field axis ($x$-axis) and to the scan axis ($y$-axis). The phase pattern on the SLM needs to be aligned to this coordinate system carefully. A final alignment step is implemented, after mechanical alignment of the SLM in the illumination light path, by means of an affine coordinate transformation between the coordinates of the SLM pixel grid and the pupil coordinates in the back focal plane of the objective lens. In this way three types of misalignment can be corrected. First, the in-plane rotation can be corrected, so that the scan lines are imaged on the sensor parallel to the field axis defined by the pixel grid. Second, the SLM is slightly tilted with respect to the optical axis to separate the incoming and outgoing beam by $11{{}^{\circ }}$, see Fig. 5. Projection of the tilted sensor plane to the plane perpendicular to the optical axis, distorts the circular beam profile to an elliptical one. This effect may be compensated by an anisotropic stretch in the affine correction transformation. Third, a slight misalignment of the origin of the SLM can be compensated for by a shift of the grating pattern. A small misalignment in the axial position of the SLM can be corrected by adding a small amount of defocus to the induced aberration.

The affine transformation between the coordinates in the SLM plane $\vec {\sigma }$ and the pupil coordinates $\vec {\rho }$ is given by:

$$\vec{\rho} = \begin{bmatrix} \cos\theta & - \sin\theta \\ \sin\theta & \cos\theta \\ \end{bmatrix}\begin{bmatrix} M_{x} & 0 \\ 0 & M_{y} \\ \end{bmatrix}\left( \vec{\sigma} - {\vec{\sigma}}_{0} \right),$$
where the first matrix describes the rotation over an angle $\theta$ around the $z$-axis, the second matrix represents the (anisotropic) scaling due to the tilt of the SLM in the light path and due to the overall magnification from the SLM plane to the back focal plane of the objective, and ${\vec {\sigma }}_{0}$ is the position offset with respect to the optical axis.

The following aberration profile is fed to the SLM:

$$W = g\left( \textrm{mod}\left\lbrack \frac{1}{\lambda}A\frac{\sigma_{x}}{R_{s}} \right\rbrack \right) + g\left( \textrm{mod}\left\lbrack \frac{1}{\lambda}A\frac{\sigma_{y}}{R_{s}} \right\rbrack \right) + \frac{1}{2}B\frac{|\vec{\sigma}|^{2}}{R_{S}^{2}},$$
where $g$ is a pattern function, $\lambda$ is the wavelength, $R_{s}$ is the radius of the illuminated area on the SLM, and $A$ and $B$ are freely configurable parameters in units of $\lambda$. The first two terms will diffract the light along the horizontal and vertical directions, respectively. The pattern function $g$ is chosen such that light is uniformly distributed over the first 9 diffraction orders (−4th to +4th order). The third term will introduce a defocus, but for nonzero $\vec {\sigma }_{0}$ it also introduces a lateral shift [27]. The measurement is repeated using a range of 41 values of $B$ spaced with $\Delta B = 90\;\textrm{m}\lambda $. This aberration profile results in a grid of 9$\times$9 spots. The position of spot $(k,\;l,\;m)$, with $k = - 4,\ldots ,4$, $l = - 4,\ldots ,4$, and $m = - 20,\ldots ,20$ in the front focal plane is then given by:
$${\vec{x}}_{klm} = \frac{F_{0}}{R_{s}} \begin{bmatrix} \cos\theta & - \sin\theta \\ \sin\theta & \cos\theta \\ \end{bmatrix}\begin{bmatrix} 1/M_{x} & 0 \\ 0 & 1/M_{y} \\ \end{bmatrix}\left( A\begin{bmatrix} k \\ l \\ \end{bmatrix} + m\Delta B\frac{{\vec{\sigma}}_{0}}{R_{s}} \right) + {\vec{x}}_{0},$$
where ${\vec {x}}_{0}$ is an overall offset with respect to the optical axis.

The image data and the spot positions of three out of the 41 measurements are shown in Figs. 7(a) and 7(b), respectively. The change of the spot positions as a function of $B$ with respect to their average position is shown in Fig. 7(c). A linear least squares fit is used to fit the model of Eq. (37) to the measured spot positions. The fit yields the values $\theta = 0.8{{}^{\circ }}, M_{x} = 1.014, M_{y} = 1.015, {\vec {\sigma }}_{0}/R_{s} = ( - 0.03, 0,08)$. These values are small, as is expected of a mechanically well aligned system, but in need of correction nevertheless.

 figure: Fig. 7.

Fig. 7. Illustration of the alignment calibration using a spot grid. (a-c) image data for three different $B$ values. (d-f) The found spot positions (blue circles) and the fitted model (red crosses). (g-h) The change in the $x$ and $y$ spot coordinate with respect to their average value as a function of $B$ (blue) and the fitted change in position (red).

Download Full Size | PDF

3.6 Correction of aberrations in the illumination light path

The designed illumination PSF, comprising the set of multi-focal scan lines, will be distorted by unintended aberrations in the illumination light path. These can arise from a (coarse) non-uniformity of the SLM pixel response function, or by the components and/or misalignment of the beam expander and the relay path between the SLM and the objective back focal plane. Fortunately, the SLM that is used to introduce intentional aberrations, can also be used to remove the unintentional aberrations. To this end, an image based method is used, without the need for any additional optical components in the setup [36]. The optimization metric that is used is the peak intensity of a single imaged spot (Strehl ratio optimization). A focused spot is imaged using the thick uniformly fluorescent calibration slide, which is aligned such that the object plane lies about 10 µm deep in the material. In this way a depth averaged illumination PSF is imaged, but this is not problematic for the intended aberration retrieval. The actual peak intensity is determined from a two dimensional parabolic fit of the pixel values around the peak. The to-be corrected aberration profile is parameterized as a sum of orthogonal Zernike polynomials $Z_{nm}$ with radial index $n$ and azimuthal index $m$, where all modes with $n + |m| \leq ~6$ are taken into account, except for piston, tip and tilt. This takes primary and secondary astigmatism, coma, and spherical aberration into account, as well as primary trefoil. Figure 8 illustrates the search algorithm used to find the optimal aberration. For each Zernike mode a set of typically three to five spots are recorded with an aberration coefficient that is sequentially changed by an amount $\pm \mathrm {\Delta }A$, where the sign is set such that a local optimum in the fitted peak intensity is within the range of recorded aberration settings. A value $\mathrm {\Delta }A = 50\;\textrm{m}\lambda $ is found to properly sample the optimization metric, while only a small number of measurements is required to find the optimum. The optimum aberration value is found by parabolic fitting of the set of fitted peak intensities, where only the maximum value and its two neighbors are included. Figure 8(a) illustrates this search method for primary vertical astigmatism. In this case, four measurements where needed to find the optimum. The corresponding spots are displayed in Fig. 8(c). This procedure is repeated for all Zernike modes. The result is shown in Fig. 8(b). Confidence intervals were obtained by repeating the whole procedure ten times on different parts of the sample. Apart from the contribution of $Z_{20}$ (defocus), the total root mean squared (rms) aberration was found to be $168 \pm 20\;\textrm{m}\lambda $, composed mainly of astigmatism, and significantly higher than the diffraction limit of $72\;\textrm{m}\lambda $. Figure 8(d) show the illumination spot without any correction, after minimizing the contribution of $Z_{20}$, and after aberration correction. Clearly, correction of unintended aberrations is a necessity in our setup.

 figure: Fig. 8.

Fig. 8. (a) Example showing the measurements performed to minimize vertical astigmatism. Four measurements where required to coarsely find the optimum (blue circles). From these measurements a fine estimation of the optimum was calculated using parabolic interpolation (yellow). (b) Graph showing the contributions of the Zernike polynomials to the total aberration. The error bars indicate two standard deviations. (c) The raw image data underlying the vertical astigmatism measurement. (d) The depth averaged illumination spot without any corrections, after minimizing the contribution of $Z_{20}$, and after aberration correction. The red circles indicate one Airy unit.

Download Full Size | PDF

4. Experimental results

4.1 Characterization of the illumination pattern

The line profile of the illumination PSF is tested experimentally using the thin uniformly fluorescent test slide. The thin fluorescent layer is aligned such that the central illumination line is in focus. The sample translation stage is then used to slowly move the slide during image acquisition in order to create a small ‘motion blur’ for smoothing out sample irregularities and for spreading the impact of photo-bleaching. The measurement is repeated ten times and the resulting images are averaged for further reduction of noise and sample irregularities. According to the image formation model the expected line shapes are the convolution of the line excitation PSF at focus $m\mathrm {\Delta }z$ and the average of the emission PSF in the field direction, at focus $- m\mathrm {\Delta }z$. The difference between the imaging NA (equal to 0.75) and the line illumination NA (equal to 0.26) implies that near focus the width of the line excitation PSF is much larger than the width of the emission PSF, whereas further away from focus this is the opposite. The reason is that the relatively small depth of focus of the emission PSF results in a steep increase in the lateral width with defocus.

Figure 9(a) shows the full FOV of the image sensor with the illumination PSF, showing the successfully creation of 9 equally spaced focal lines (labeled with their diffraction order number $m)$. Figure 9(b) shows the intensities of the focal lines along the field direction. The higher order astigmatism based design ensures a substantial illumination intensity over the whole line length. The remaining decrease of power towards the edges is caused by the Gaussian beam profile at the objective pupil. In practice, a decrease of intensity at the edges of the lines to about 30 % is observed. The lines show irregularity on a small length scale resulting from laser speckle. Further, a significant decrease in intensity is visible around $x = 0$ µm for all lines, resulting from an uneven illumination of the SLM. Figure 9(c) displays the FWHM of the lines. The center lines are in focus and approach the theoretical limit of FWHM $= 0.89 \lambda /\textrm {NA}{}$. The FWHM of the outer lines ($m = - 4, - 3,3,4)$ are substantially larger, as they are recorded out of focus, in agreement with the image formation model. A slight tilt of the sample in the field direction is apparent from the fact that the lines with $m\;>\;0$ have a decreasing FWHM for larger $x$ and vice versa for the lines with $m\;<\;0$.

 figure: Fig. 9.

Fig. 9. The illumination PSF measured using a thin uniformly fluorescent test slide. (a). The full illumination PSF. The scale bar represents 100 µm. The focal lines are labeled with their diffraction order number $m$ and with aranging from blue for $m = - 4$ to yellow for $m = 4.$ The red lines indicate the width of the regions of interest (ROIs) that are used for imaging. The design line length $2l = 488$ µm is indicated by green lines. Further indicated are the SLM’s zeroth order spot (A), the first alias to the right of the central pattern (B), and higher order focal lines ($\left | m \right |\;>\;4)$ that are folded back into the central domain due to undersampling (C). These are minor artifacts that are blocked by the spatial slit filter in the digital domain. (b) The intensity distribution over the lines. The higher order astigmatism successfully prevents a zero rim intensity. A decrease of intensity is observed around $x = 0$ µm which is caused by an uneven illumination of the SLM. (c) The full width at half maximum (FWHM) of the focal lines. The center lines are in focus and approach the theoretical limit (full black line), a bit broader than the height of a ROI for four line TDI (dashed black line). The lines in (b) and (c) correspond to the diffraction orders as indicated at the left side of the figure, the width of the $x$-axis is scaled to the design length $2l$, and the red lines indicate the ROI width.

Download Full Size | PDF

The sectioning capability of the system is tested by measuring the background function as defined in Eq. (29). This is done by repeating the previous measurement for a series of axial sample stage positions. The 10 times averaging is omitted for this measurement to limit the illumination photon dose and so avoid bleaching of the sample. The measured images are integrated over the digital slit in the $y$ (scan) direction (up to 4 pixel rows) and over 50 pixels about halfway the line center and the line edge (at around $x=w/4$) in the $x$ (lateral) direction, for noise suppression. As a result, we obtain a background signal as a function of the stage position $z$ for all 9 lines, see Fig. 10. For the first result, in Fig. 10(a), a single row on the sensor is used as confocal detector. The background function has an average full width at half maximum $\textrm {FWHM}{}_z=8.4$ µm. In Fig. 10(b), the result is given the case of a four row confocal detector, which slightly increases the $\textrm {FWHM}{}_z$ to 9.0 µm on average. The theoretical values for the response to a uniformly fluorescent thin slide are given in Fig. 10(e). Qualitatively similar curves are found, but less wide than the experimental curves. The $\textrm {FWHM}{}_z$ is found to be 4.2 µm. The background measured at the bottom scan plane from a source located at the top scan plane is about 20 %, where from the simulation a reduction to about 10 % would have been expected. This result is based on a numerical simulation that computes the illumination PSF $H_{ill}$, given a defocus following from a stage position that is $z_s$ out-of-focus, according to Eq. (22). Then, the intensity at the sensor $I$ is calculated by convolution of $H_{ill}$ with the imaging PSF corrected for a stage defocus of $z_s$, see Eq. (23). The simulated intensity $I$ is saved to disk as an image file and analyzed in the same way as the experimental results. The discrepancy between the width of the measured peak in the background function and the theoretical expectation may be attributed to misalignment, the finite thickness of the layer, and residual aberrations, mainly spherical aberration. The use of scalar diffraction for computing the PSFs instead of more realistic vector PSF models could also lead to a predicted background function peak width that is too optimistic.

 figure: Fig. 10.

Fig. 10. Confocal detection of a uniform fluorescent plane. The top row (a, b, c) shows the signal as a function of the sample position in air, for the 9 illumination lines. The bottom row (d, e, f) shows the corresponding axial FWHM. The focal lines are labeled with their diffraction order number $m$ and with aranging from blue for $m = - 4$ to yellow for $m = 4.$ (a, b) The measured signal, while a single row on the sensor is used as a confocal detection. (c, d), The measured response for a four-row confocal detection. (e, f) The simulated response for a four-row confocal detection.

Download Full Size | PDF

The illumination PSF is further tested for uniform spacing and for equidistant foci. To this end, the thick uniformly fluorescent test slide is aligned such that the focal point of the objective is about 10 µm deep in the material. A through-focus stack is created by repeatedly adding a defocus aberration $W$ to the SLM and capturing an image. According to the image formation model the expected line shape is the 2D convolution of the through-focus line excitation PSF and the average of the through-focus emission PSF. Because of the much smaller illumination NA this implies that the measured through-focus line profiles are largely determined by the through-focus line excitation profiles, and hardly by the through-focus emission PSF. Consequently, by using an aberration $W = z\sqrt {n^{2} - \rho ^{2}\textrm {NA}{}^{2}}$, the illumination PSF is sampled at a depth $z$.

Figure 11(a) shows the measured through focus stacks for an illumination PSF of the 9 beamlets, averaged across the $x$ (field) direction. To find the focus, for every beamlet the position of the maximum intensity is determined. Figure 11(b) displays these positions in the $(y,\;z)$ plane. The positions have an equal spacing in the $y$ direction of 54.5 µm or 200 pixels on the sensor with a deviation less than 0.14 µm (0.5 pixel). In the $z$ direction the sub beams do have the desired tilted focus. The graph includes a linear fit through the focal points, which overlaps with the tilted object plane with a deviation less than 0.2 µm. However, on top of the linear focus shift, we observe an apparent field curvature leading to a defocus of up to 1.8 µm at the edges. This leads to a slight axial misalignment of the illumination PSF compared to the reference axial imaging depth defined by the imaging PSF. The imaging depth will, however, only be slightly biased as the NA of the illumination is lower than the NA of the imaging path, and therefore the latter will dominate the imaging depth.

 figure: Fig. 11.

Fig. 11. Assessment of illumination PSF in a thick fluorescent material. (a) Through focus stack averaged along the $x$ direction. The red crosses mark the focal points. Thebars below the images indicate the intensity value mapping that is adjusted to cover the full dynamic range in each image. (b) The focal points plotted in the $(y,\;z)$ plane (blue circles) including a linear fit through the focal points (blue) which overlaps the object plane (red dashed curve).

Download Full Size | PDF

4.2 Scan results

We present three scan result as proof of concept. First, a 9 layer multi-focal whole slide scan is created of the FISH labeled TMA tissue section. Figure 12(a) shows a maximum intensity projection of the whole slide scan. The image is composed of thirteen scan lanes of 20,000 lines with 1536 pixels width. After resampling this results in a 3600 Mpixel volumetric image ($19,968\times 18,794\times 9$ pixels, size $5,790\times 5,450\times 11$ µm). For this acquisition, an exposure time of 1 ms and four TDI lines are used. Figures 12(b)–12(f) provide closer details of the scan and show that zooming in to the cellular level for inspecting tissue structure in Fig. 12(c) and resolving individual FISH spots in Fig. 12(f) is possible.

 figure: Fig. 12.

Fig. 12. A 400 Mpixel maximum intensity projection of a 9 layer multi-focal scan of a tissue micro array (TMA) section labeled using FISH. Every figure zooms in a factor of five, such that the images are shown over a factor of 125 different length scales. The scan is made using an exposure time of 1 ms and four TDI lines. A gamma of 0.7 is used for an enhanced dynamic range. This image is also shown over 100 length scales in Visualization 1. (a) Whole slide scan. The white boxes indicate two regions that displayed in more detail. (b, d) 5$\times$ enlarged. (c, e) 25$\times$ enlarged. (f) 125$\times$ enlarged.

Download Full Size | PDF

The second scan result displayed in Fig. 13 shows the multi-focal imaging capability of the system. The same FISH labeled tissue is scanned using a single line detector (i.e. no TDI) for better confocality. Here, an exposure time of 10 ms is used. Figure 13(a) shows a frame of a 3D visualization of the $z$-stack, created using the image analysis software Icy [37]. FISH spots can be observed to have different $z$-positions. This is further illustrated by the color coded maximum intensity projection shown in Fig. 13(b). In this image, the luminescence of each pixel corresponds to the maximum intensity throughout the layers, while the color corresponds to the layer of maximum intensity. Most spots color orange, indicating a depth of about 3 µm. Some are observed to lay deeper in the sample, ranging from −1 µm (green) to −3 µm.

 figure: Fig. 13.

Fig. 13. 3D visualizations of a 9 layer multi-focal scan of a tissue micro array (TMA) section labeled using FISH. The scan is made using an exposure time of 10 ms and a single line detector. (a) A frame from a 3D visualization using the Icy software [37] (see Visualization 2). The white line grid has a 10 µm spacing. (b) A maximum intensity projection, where pixels have a assigned according to the layer of maximum intensity, ranging from −5.2 µm (blue) to 5.2 µm (yellow). The FISH probes can be observed to have different positions in $z$. The scale bar represents 10 µm.

Download Full Size | PDF

The third result is a scan of the stage 3 human rectum cancer sample with immunofluorescence staining (10 ms exposure time and using a single pixel row as digital slit). Figure 14 shows the central layer of a single lane scan. Fig. 14(b) gives a detailed view which provides a good benchmark for the resolution and contrast delivered by the system.

 figure: Fig. 14.

Fig. 14. A single lane, 9 layer multi-focal scan of an immunofluorescently labeled human rectum tissue section. The scan is made using an exposure time of 10 ms and a single line digital spatial filter. (a) Central layer of the scan, showing the full width of a scan lane. The white box indicates the area that is showed in detail. (b) Detailed view. The white box marks the area displayed in (c-k), which provides a through focus stack of all layers, demonstrating the optical sectioning capability. Although this is a thick tissue section, a high contrast is observed in the central layers.

Download Full Size | PDF

For all scanned samples the reduction of out-of-focus background light by the digital slit pinhole is moderate, in agreement with the measured background function with $\textrm {FWHM}{}_z$ of around 9 µm, and the thickness of the samples of around 5 µm.

4.3 Throughput and resolution

The throughput of the current test setup is $T = N_{z}N_{x}f$, with $N_{z} = 9$ the number of focus layers, $N_{x} = 1536$ the line length in pixels, and $f = 21$ lines/s the line rate, leading to a modest $T = 0.29$ Mpixels/s. This number is primarily limited by the practical constraints of the image sensor, as the actual exposure time in TDI-mode was just 1 ms, implying a potential gain in throughput with a factor of about 50. A second practical limitation is in the illumination light path efficiency. The main losses were measured to be the coupling into the fiber (35 % loss), the polarizer (35 % loss), the iris that cuts the beam to about one sigma (39 % loss) and the SLM (48 % loss), accumulating to a total illumination light path efficiency of 14 %. These losses are not fundamental and can be improved upon, for example by better alignment of the half wave plate to the polarizer; using free space optics instead of a fiber; a beam shaping optical element to reshape the Gaussian beam into a top hat beam [38]; and by using a dedicated diffractive structure etched in quartz or embossed in a polymer instead of the SLM. A third limitation is the reduced detection efficiency of the camera for light incident at an angle due to the micro lens array (47 % loss). Overcoming these practical hurdles can improve throughput with a factor ranging between 200 and 600 without affecting the SNR. The exposure time could be reduced by an increase in laser power to a decent but not excessive 500 mW. This would further enhance the throughput with a factor of 5. As a result, the proposed optical architecture is estimated to support image acquisitions with a throughput on the order of several hundred Mpixel/s. In the current setup, that would correspond to a scanning speed on the order of 10 mm/s. Note that mechanics do not pose practical constrains on the throughput as the only movement in the system is the linear translation of the sample.

The fundamental limitation to throughput of the proposed scanning architecture is the minimum SNR that is required [39]. The SNR is dominated by shot noise, and therefore the achievable throughput is governed by the minimum amount of photons that must be captured. The maximum line rate $f_{max}$ is proportional to $\eta KP/N_{\min }$, with $\eta$ the overall photon efficiency of the illumination and imaging light path and detector, $K$ the number of TDI-lines, $P$ the laser power, and $N_{\min }$ the minimum photon count for a reference fluorescent object.

The relation between the throughput and SNR that is currently achieved is further assessed on the basis of the scan result of the whole slide scan of the FISH slide, see Fig. 12. This $5\times 5$ mm tissue section is scanned in 12 lanes of about 17,000 lines, which gives a total camera time (exposure time times number of lines) of less than 3.5 min. We have measured the peak signal to noise ratio (pSNR) and the peak signal to background ratio (pSBR) in a region of 445 µm by 558 µm with a dense distribution of FISH dots. A maximum intensity projection of this region is shown in Fig. 15(a). A zoomed-in view of three FISH spots are shown in Figs. 15(b)–15(d), where we selected the layer with best focus. A classification is performed in which each pixel is classified as being part of the background, part of a spot or neither of both. The classification is done on the 2D maximum intensity projection. The 20th percentile lowest values displayed in Fig. 15(e) are classified as background. An average background level of 139±7 e is found. The pixels with the 0.5th percentile highest are classified as part of a spot or cluster of spots. These pixels are segmented based on 4-connectivity. Small segments are rejected by requiring the shape to span at least three columns and three rows. The peak intensity in every segment was determined by first selecting the layer with best focus and then by interpolating the peak intensity using a two dimensional parabolic fit of the values surrounding the pixel with the maximum intensity to the the model $I(x,\;y) = w_x(x-x_c)^{2} + w_y(y-y_c)^{2} + I_p, \label {eq:2d-par-Ip}$ where $I$ is the image intensity, $w_x$, $w_y$, $x_c$ and $y_c$ are fit parameters and $I_P$ provides the peak intensity. In total, 109 segments are found with peak values ranging from 908 to 1739 e. For every segment the pSBR is calculated by dividing the peak intensity by the average background level. Figure 15(f) shows a histogram of the obtained values. An average pSBR of 8.4 (6.5 - 12.0) is found. The pSNR for every segment is approximated as the square root of the peak intensity, as the noise is dominated by shot noise. Figure 15(g) shows a histogram of the obtained pSNR values. An average of 34 (27 - 38) is found. The pSNR and pSBR values that are found are sufficient for state-of-the-art spot segmentation algorithms, which require an SNR level around 5 [40]. This implies that in principle the scan speed could have been 7 times faster.

 figure: Fig. 15.

Fig. 15. Determination of the peak signal to noise ratio (pSNR), the peak signal to background ratio (pSBR), and the lateral and axial resolution expressed using the full width at half maximum (FWHM). (a) Maximum intensity projection of the image data of a scan of the FISH slide made using 1 ms exposure time and a four-line TDI acquisition. The scale bar indicates 100 µm. A gamma correction of 0.7 is applied. Three locations are highlighted. (b-d) A zoomed in view to the highlighted areas. The layer with best focus is displayed. The spots location are given including their peak intensity in photo electrons (e). The scale bar indicates 2 µm. The yellow ellipses have a horizontal and vertical diameter equal to the FWHM along the and direction respectively. (e) The image data was classified as background (black) and non-background values (green) or part of a spot. In total, 109 spots where segmented and their location is indicated with a yellow dot. (f-j) Histograms of the pSNR, pSBR and the FWHM along the $x$, and $y$ direction. The red line indicates the mean and the red dashed lines the 5th and 95th percentile respectively. In (h-j) a black line is included indicating the FWHM of the imaging PSF.

Download Full Size | PDF

The lateral and axial resolution of the system are assessed based on the same dataset. Close to the maximum, the spot intensity is well approximated by a two dimensional parabola. Consequently, the lateral full width at half maximum (FWHM) is estimated from the fit parameters $w_x,\;w_y$ according to $\textrm {FWHM}_{(x,\;y)} = 2\sqrt {I_{p}/2w_{(x,\;y)}}$. Histograms summarizing the results are shown in Figs. 15(h)–15(j). The lower bounds of the distributions provide the best indication of the resolution, as the segments include both single spots and spot clusters. The 5th-percentile is 0.50±0.13 µm in the $x$ direction (along the illumination line). In the $y$ direction (orthogonal to the illumination line i.e. the scan direction) a value of 0.53±0.14 µm is found. The FWHM of the emission PSF according to the scalar diffraction based Airy distribution convolved with the pixel size is given by $0.57\lambda /\textrm {NA}{} = 0.43$ µm for an emission wavelength $\lambda =563$ nm, and is included in Figs. 15(h)–15(j) as black lines for comparison. The obtained values agree reasonably well with this theoretical value. The axial resolution was assessed in a similar fashion. For every segment, the maximum intensity in every layer $I_{z}(z)$ is determined. The values around the maximum are fitted with the parabolic model $I_{z}\left ( z \right ) = w_{z}\left ( z - z_{c} \right )^{2} + I_{\textrm {pz}}$. Then, $\textrm {FWH}M_{z} = 2\sqrt {I_{\textrm {pz}}/2w_{z}}$. Figure 15(j) shows a histogram of the obtained result. The FWHM of the emission PSF is given by $1.61 \lambda /(n-\sqrt {n^{2}-NA^{2}}) = 4.52$ µm and is included as a black line. The 5th-percentile of the found distribution is 3.35±0.10 µm, which exceeds the theoretical value. Factors impacting this analysis are the limited number of focal layers and the residual non-uniformity of the illumination intensity between the the focal layers.

5. Discussion

In summary, we presented a multi-line scanner for multi-focal fluorescence image acquisitions in a single scan. The sample is scanned with respect to a mechanically fixed optical setup, thereby providing simple means for an unlimited field of view. The optical architecture of the scanner is based on illumination PSF engineering. We used a design with diffractive optics for generating a set of parallel scan lines, focused at equidistant focal planes, in combination with a tilted sensor, for capturing the emitted fluorescence from the scan lines in parallel. An important new element in the design is the use of higher order astigmatism to improve the uniformity of peak intensity and line width along the scan lines. The approach is suitable for scanning large area tissue samples, as needed for imaging e.g. immunofluorescently labeled tissue samples. The ultimate goal for large area FISH imaging is to improve the statistical reliability in diagnosis, by increasing the number of cells tested and by correlating the resulting gene counts to the heterogeneous tissue structure.

The current setup is limited in the overall axial range of the 3D scan to a value between 10 and 15 µm. A larger axial range could for example be obtained by replacing the 20$\times$ objective with a 10$\times$ objective, leading to a fourfold increase in axial range. This, however, comes at the expense of lateral resolution, as the decrease in magnification is usually attended by a decrease in NA as well. Such an intended increase in axial range would also entail producing more than the current 9 parallel scan lines.

A major technical improvement of the current setup would be the use of an SLM with more pixels, e.g. $1920 \times 1152$ instead of the current $512 \times 512$. The enhanced space-bandwidth product could enable focal lines with an illumination NA matching the detection NA, equal to 0.75, instead of the current value 0.26, while maintaining or increasing the length of the scan lines of close to 0.5 mm. The main impact would be the improved sectioning capability, where our PSF simulations indicate that a peak in the background function with a width equal to 1.5 µm can be realized.

A next step for the proposed multi-line confocal scanner is the extension to multi-colour imaging. This can be realized using time multiplexing or space multiplexing. For time multiplexing the system cycles through the different excitation wavelengths. For each excitation wavelength a set of multi-focal scan lines must be generated, either via a fast switching SLM or by a setup with a modulator to select the wavelength and a dichroic beam splitter architecture wherein each color branch has a separate diffractive optical element, designed for the particular wavelength. For space multiplexing, the different sets of multi-focal scan lines for the different excitation wavelengths are slightly displaced with respect to each other in the direction of the tilted plane imaged onto the sensor. As a consequence, the different color channels can be imaged in parallel, making more economical use of the sensor area. This approach does not require switching of the illumination and is compatible with a rolling shutter read-out.

Another extension that is envisioned is the application of techniques to improve the resolution. In particular, the Image Scanning Microscopy (ISM) method and the closely related pixel reassignment and rescan methods [4144]. In fact, a modification of the TDI-approach we use can already deliver an ISM type of resolution improvement in the scan direction. This would require oversampling in the scan direction in combination with a digital stretch of the captured images in the scan direction before the step of adding the images acquired in subsequent frames. Although relatively straightforward, this would have the drawback of an anisotropic resolution, as the imaging in the field direction, along the scan lines, remains unaffected. A more isotropic improvement in resolution would require a modification of the multi-focal line illumination to a multi-focal multi-spot illumination, i.e. an illumination PSF which is spatially varying in the direction orthogonal to the scan direction as well. Such methods could overcome the drawback of the current scanning approach that it is not easily compatible with immersion objective lenses for high-resolution imaging.

Funding

Nederlandse Organisatie voor Wetenschappelijk Onderzoek (Netherlands Organization for Scientific Research - Applied and Engineering Sciences, NWO-TTW) (Project 13892).

Acknowledgments

We wish to thank Bas Hulsken, Rik Kneepkens, Jelte Peter Vink, Mathijs Rem, Gert-Jan van den Braak and Aslihan Arslan Carisey of Philips Digital & Computational Pathology, Hein Sleddens of the Erasmus University Medical Center and Hans Morreau of the Leiden University Medical Center for valuable research input.

Disclosures

The authors declare that there are no conflicts of interest related to this article.

References

1. F. Zanella, J. B. Lorens, and W. Link, “High content screening: seeing is believing,” Trends Biotechnol. 28(5), 237–245 (2010). [CrossRef]  

2. G. Bueno, M. M. Fernández-Carrobles, O. Deniz, and M. García-Rojo, “New Trends of Emerging Technologies in Digital Pathology,” Pathobiology 83(2-3), 61–69 (2016). [CrossRef]  

3. T. Dorval, B. Chanrion, M.-E. Cattin, and J. P. Stephan, “Filling the drug discovery gap: is high-content screening the missing link?” Curr. Opin. Pharmacol. 42, 40–45 (2018). [CrossRef]  

4. J. B. Pawley, Handbook Of Biological Confocal Microscopy (Springer, 2006).

5. L. Tian, Z. Liu, L.-H. Yeh, M. Chen, J. Zhong, and L. Waller, “Computational illumination for high-speed in vitro Fourier ptychographic microscopy,” Optica 2(10), 904–911 (2015). [CrossRef]  

6. E. Wang, C. M. Babbey, and K. W. Dunn, “Performance comparison between the high-speed Yokogawa spinning disc confocal system and single-point scanning confocal systems,” J. Microsc. 218(2), 148–159 (2005). [CrossRef]  

7. T. Tanaami, S. Otsuki, N. Tomosada, Y. Kosugi, M. Shimizu, and H. Ishida, “High-speed 1-frame/ms scanning confocal microscope with a microlens and Nipkow disks,” Appl. Opt. 41(22), 4704–4708 (2002). [CrossRef]  

8. H. L. Fu, J. L. Mueller, M. P. Javid, J. K. Mito, D. G. Kirsch, N. Ramanujam, and J. Q. Brown, “Optimization of a Widefield Structured Illumination Microscope for Non-Destructive Assessment and Quantification of Nuclear Features in Tumor Margins of a Primary Mouse Model of Sarcoma,” PLoS One 8(7), e68868 (2013). [CrossRef]  

9. T. C. Schlichenmeyer, M. Wang, K. N. Elfer, and J. Q. Brown, “Video-rate structured illumination microscopy for high-throughput imaging of large tissue areas,” Biomed. Opt. Express 5(2), 366–377 (2014). [CrossRef]  

10. P.-H. Wang, V. R. Singh, J.-M. Wong, K.-B. Sung, and Y. Luo, “Non-axial-scanning multifocal confocal microscopy with multiplexed volume holographic gratings,” Opt. Lett. 42(2), 346–349 (2017). [CrossRef]  

11. A. Jesacher, S. Bernet, and M. Ritsch-Marte, “Colored point spread function engineering for parallel confocal microscopy,” Opt. Express 24(24), 27395–27402 (2016). [CrossRef]  

12. L. Li, Q. Zhou, T. C. Voss, K. L. Quick, and D. V. LaBarbera, “High-throughput imaging: Focusing in on drug discovery in 3D,” Methods 96, 97–102 (2016). [CrossRef]  

13. R. Wolleschensky, B. Zimmermann, and M. Kempe, “High-speed confocal fluorescence imaging with a novel line scanning microscope,” J. Biomed. Opt. 11(6), 064011 (2006). [CrossRef]  

14. K.-B. Im, S. Han, H. Park, D. Kim, and B.-M. Kim, “Simple high-speed confocal line-scanning microscope,” Opt. Express 13(13), 5151–5156 (2005). [CrossRef]  

15. C. Sheppard and X. Mao, “Confocal Microscopes with Slit Apertures,” J. Mod. Opt. 35(7), 1169–1185 (1988). [CrossRef]  

16. C.-M. Chia, H.-C. Wang, J. A. Yeh, D. Bhattacharya, and Y. Luo, “Multiplexed holographic non-axial-scanning slit confocal fluorescence microscopy,” Opt. Express 26(11), 14288–14294 (2018). [CrossRef]  

17. S. Abeytunge, Y. Li, B. Larson, R. Toledo-Crow, and M. Rajadhyaksha, “Rapid confocal imaging of large areas of excised tissue with strip mosaicing,” J. Biomed. Opt. 16(5), 050504 (2011). [CrossRef]  

18. S. M. Shakeri, B. Hulsken, L. J. van Vliet, and S. Stallinga, “Optical quality assessment of whole slide imaging systems for digital pathology,” Opt. Express 23(2), 1319–1336 (2015). [CrossRef]  

19. T. Yang, T. Zheng, Z. Shang, X. Wang, X. Lv, J. Yuan, and S. Zeng, “Rapid imaging of large tissues using high-resolution stage-scanning microscopy,” Biomed. Opt. Express 6(5), 1867–1875 (2015). [CrossRef]  

20. P. J. Cutler, M. D. Malik, S. Liu, J. M. Byars, D. S. Lidke, and K. A. Lidke, “Multi-Color Quantum Dot Tracking Using a High-Speed Hyperspectral Line-Scanning Microscope,” PLoS One 8(5), e64320 (2013). [CrossRef]  

21. M. E. Gehm, M. S. Kim, C. Fernandez, and D. J. Brady, “High-throughput, multiplexed pushbroom hyperspectral microscopy,” Opt. Express 16(15), 11032–11043 (2008). [CrossRef]  

22. B. Hulsken, “Method for simultaneous capture of image data at multiple depths of a sample,” US Patent No. 9,910,258 B2 (2018).

23. J. Broeken, B. Rieger, and S. Stallinga, “Simultaneous measurement of position and color of single fluorescent emitters using diffractive optics,” Opt. Lett. 39(11), 3352–3355 (2014). [CrossRef]  

24. A. Greengard, Y. Y. Schechner, and R. Piestun, “Depth from diffracted rotation,” Opt. Lett. 31(2), 181–183 (2006). [CrossRef]  

25. Y. Shechtman, S. J. Sahl, A. S. Backer, and W. E. Moerner, “Optimal Point Spread Function Design for 3D Imaging,” Phys. Rev. Lett. 113(13), 133902 (2014). [CrossRef]  

26. Y. Shechtman, L. E. Weiss, A. S. Backer, M. Y. Lee, and W. E. Moerner, “Multicolour localization microscopy by point-spread-function engineering,” Nat. Photonics 10(9), 590–594 (2016). [CrossRef]  

27. M. Siemons, C. N. Hulleman, R. Ø. Thorsen, C. S. Smith, and S. Stallinga, “High precision wavefront control in point spread function engineering for single emitter localization,” Opt. Express 26(7), 8397–8416 (2018). [CrossRef]  

28. C. Smith, M. Huisman, M. Siemons, D. Grünwald, and S. Stallinga, “Simultaneous measurement of emission color and 3D position of single molecules,” Opt. Express 24(5), 4996–5013 (2016). [CrossRef]  

29. B.-C. Chen, W. R. Legant, K. Wang, L. Shao, D. E. Milkie, M. W. Davidson, C. Janetopoulos, X. S. Wu, J. A. Hammer, Z. Liu, B. P. English, Y. Mimori-Kiyosue, D. P. Romero, A. T. Ritter, J. Lippincott-Schwartz, L. Fritz-Laylin, R. D. Mullins, D. M. Mitchell, J. N. Bembenek, A.-C. Reymann, R. Böhme, S. W. Grill, J. T. Wang, G. Seydoux, U. S. Tulu, D. P. Kiehart, and E. Betzig, “Lattice light-sheet microscopy: Imaging molecules to embryos at high spatiotemporal resolution,” Science 346(6208), 1257998 (2014). [CrossRef]  

30. B.-J. Chang, M. Kittisopikul, K. M. Dean, P. Roudot, E. S. Welf, and R. Fiolka, “Universal light-sheet generation with field synthesis,” Nat. Methods 16(3), 235–238 (2019). [CrossRef]  

31. B. R. W. Gerchberg and W. O. Saxton, “A Practical Algorithm for the Determination of Phase from Image and Diffraction Plane Pictures,” Optik 35, 237–246 (1972).

32. S. Stallinga, “Finite conjugate spherical aberration compensation in high numerical-aperture optical disc readout,” Appl. Opt. 44(34), 7307–7312 (2005). [CrossRef]  

33. S. M. Shakeri, “Whole Slide Imaging Systems for Digital Pathology,” Dissertation, Delft University of Technology (2017).

34. H. Netten, L. J. van Vliet, F. R. Boddeke, P. de Jong, and I. T. Young, “A fast scanner for fluorescence microscopy using a 2-D CCD and time delayed integration,” Bioimaging 3(2), 102 (1995). [CrossRef]  

35. E. Ronzitti, M. Guillon, V. de Sars, and V. Emiliani, “LCoS nematic SLM characterization and modeling for diffraction efficiency optimization, zero and ghost orders suppression,” Opt. Express 20(16), 17843–17855 (2012). [CrossRef]  

36. M. J. Booth, “Wavefront sensorless adaptive optics for large aberrations,” Opt. Lett. 32(1), 5–7 (2007). [CrossRef]  

37. F. de Chaumont, S. Dallongeville, N. Chenouard, N. Hervé, S. Pop, T. Provoost, V. Meas-Yedid, P. Pankajakshan, T. Lecomte, Y. Le Montagner, T. Lagache, A. Dufour, and J.-C. Olivo-Marin, “Icy: an open bioimage informatics platform for extended reproducible research,” Nat. Methods 9(7), 690–696 (2012). [CrossRef]  

38. I. Khaw, B. Croop, J. Tang, A. Möhl, U. Fuchs, and K. Y. Han, “Flat-field illumination for quantitative fluorescence imaging,” Opt. Express 26(12), 15276–15288 (2018). [CrossRef]  

39. L. van der Graaff, F. Boyaval, L. van Vliet, and S. Stallinga, “Fluorescence imaging for whole slide scanning using LED-based color sequential illumination,” Proc. SPIE 10679, 106790D (2018). [CrossRef]  

40. I. Smal, M. Loog, W. Niessen, and E. Meijering, “Quantitative Comparison of Spot Detection Methods in Fluorescence Microscopy,” IEEE Trans. Med. Imaging 29(2), 282–301 (2010). [CrossRef]  

41. C. B. Müller and J. Enderlein, “Image Scanning Microscopy,” Phys. Rev. Lett. 104(19), 198101 (2010). [CrossRef]  

42. A. G. York, P. Chandris, D. D. Nogare, J. Head, P. Wawrzusin, R. S. Fischer, A. Chitnis, and H. Shroff, “Instant super-resolution imaging in live cells and embryos via analog image processing,” Nat. Methods 10(11), 1122–1126 (2013). [CrossRef]  

43. S. Roth, C. J. Sheppard, K. Wicker, and R. Heintzmann, “Optical photon reassignment microscopy (OPRA),” Opt. Nanoscopy 2(1), 5 (2013). [CrossRef]  

44. G. M. De Luca, R. M. Breedijk, R. A. Brandt, C. H. Zeelenberg, B. E. de Jong, W. Timmermans, L. N. Azar, R. A. Hoebe, S. Stallinga, and E. M. Manders, “Re-scan confocal microscopy: scanning twice for better resolution,” Biomed. Opt. Express 4(11), 2644–2656 (2013). [CrossRef]  

Supplementary Material (2)

NameDescription
Visualization 1       A 400 Mpixel maximum intensity projection of a 9 layer multi-focal scan of a tissue micro array (TMA) section labeled using FISH, shown over 100 length scales.
Visualization 2       3D visualisations of a 9 layer multi-focal scan of a tissue micro array (TMA) section labeled using FISH.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (15)

Fig. 1.
Fig. 1. Schematic side view of the architecture for a multi-line fluorescence scanning microscope for multi-focal imaging with an unlimited field of view. For clarity we illustrate the concept using only two focal layers. The imaging path is shown in green, the illumination path is shown in blue. A diffractive optical element (DOE) is used for generating a set of parallel scan lines, focused at equidistant planes. Note that in this side view, the lines are orthogonal to the drawing. (a) The tube lens and objective lens form a telecentric optical system such that the tilted sensor has a tilted conjugate plane in object space. The rows on the area sensor conjugate to the line foci are used to capture the image data from the different focal slices scanned by the line foci simultaneously. (b) The sample is scanned in a continuous ‘push broom’ scanning fashion to obtain a multi-focal image with a field of view that is in principle unlimited.
Fig. 2.
Fig. 2. Demonstration of grating with tilt and defocus. (a) The zone function with $\Delta y = 10\lambda$ and $\Delta{z} = 10\lambda$. These numbers are chosen for demonstration purpose, and do not represent a realistic scenario. (b) The profile function optimized for 9 lines. (c) The resulting distribution of power over the diffraction orders. (d) The corresponding aberration function.
Fig. 3.
Fig. 3. Illustration of the proposed geometry for the aspherical cylindrical lens design. (a) In the back focal plane, we have an circular aperture which is illuminated by a beam profile $A\left ( \rho _{x},\rho _{y} \right )$. We introduce an astigmatic aberration function $W$ such that every strip in the back focal plane there is a corresponding strip in the front focal plane (red boxes in all subfigures). (b) Side view of the optical system, with the back focal plane (bfp), the objective lens (obj) and the front focal plane (ffp). (c) The strips in the front focal plane have a non-constant width $dx$. By changing the shape of the $W$, $dx$ can be modified to obtain a uniform distribution of power.
Fig. 4.
Fig. 4. Calculated intensities in the focal plane using 2D scalar diffraction. A value $l = 256 \lambda /\textrm {NA}{}$ was used. The aperture is assumed to be circular. The red lines indicate one airy unit. (a) A line created using a parabolic aberration function. (b) A line created using an aberration function with higher order astigmatism optimized for uniformity conform Eq. (17). The yellow boxes correspond to the detailed images given in c-e. (c) Close to the edge of the line profile ringing is present. (d) This ringing has disappeared around $x = - 220 \lambda /\textrm {NA}{}$. Here the line width is within one Airy unit. (e) In the center of the line pattern, the line has the smallest line width. The peak is constant over the line.
Fig. 5.
Fig. 5. Schematic illustration of the multi-focal line confocal scanner prototype. The polarization of the laser beam is aligned to the SLM using a half wave plate (HWP) and filtered using a polarizer (POL). The beam is expanded and projected onto the SLM. The diffracted light is coupled into the objective lens using a filter cube consisting of an excitation filter (EX), emission filter (EM) and a dichroic mirror (DIC) and imaged in the back focal plane of the objective. The fluorescence light is imaged onto the CMOS sensor using the tube lens. Focal lengths $F$ are given in mm. (b). The measured impact on the photon detection efficiency of tilting the camera with respect to the optical axis. For a tilt of 20° the detection efficiency is reduced with 53 % with respect to orthogonal incidence.
Fig. 6.
Fig. 6. Results of the image based calibration of the SLM response curve. (a) Measured ratio between the first and zeroth diffraction order as a function of the gray levels $\epsilon _{1}$ and $\epsilon _{2}$ of the binary grating structure. (b) Modeled diffraction order ratio for the found fit parameters. (c) The induced phase change as a function of the digital gray level. The red boxes indicate the gray values that are used to create the lookup table and correspond to a $2\pi$ phase change.
Fig. 7.
Fig. 7. Illustration of the alignment calibration using a spot grid. (a-c) image data for three different $B$ values. (d-f) The found spot positions (blue circles) and the fitted model (red crosses). (g-h) The change in the $x$ and $y$ spot coordinate with respect to their average value as a function of $B$ (blue) and the fitted change in position (red).
Fig. 8.
Fig. 8. (a) Example showing the measurements performed to minimize vertical astigmatism. Four measurements where required to coarsely find the optimum (blue circles). From these measurements a fine estimation of the optimum was calculated using parabolic interpolation (yellow). (b) Graph showing the contributions of the Zernike polynomials to the total aberration. The error bars indicate two standard deviations. (c) The raw image data underlying the vertical astigmatism measurement. (d) The depth averaged illumination spot without any corrections, after minimizing the contribution of $Z_{20}$, and after aberration correction. The red circles indicate one Airy unit.
Fig. 9.
Fig. 9. The illumination PSF measured using a thin uniformly fluorescent test slide. (a). The full illumination PSF. The scale bar represents 100 µm. The focal lines are labeled with their diffraction order number $m$ and with aranging from blue for $m = - 4$ to yellow for $m = 4.$ The red lines indicate the width of the regions of interest (ROIs) that are used for imaging. The design line length $2l = 488$ µm is indicated by green lines. Further indicated are the SLM’s zeroth order spot (A), the first alias to the right of the central pattern (B), and higher order focal lines ($\left | m \right |\;>\;4)$ that are folded back into the central domain due to undersampling (C). These are minor artifacts that are blocked by the spatial slit filter in the digital domain. (b) The intensity distribution over the lines. The higher order astigmatism successfully prevents a zero rim intensity. A decrease of intensity is observed around $x = 0$ µm which is caused by an uneven illumination of the SLM. (c) The full width at half maximum (FWHM) of the focal lines. The center lines are in focus and approach the theoretical limit (full black line), a bit broader than the height of a ROI for four line TDI (dashed black line). The lines in (b) and (c) correspond to the diffraction orders as indicated at the left side of the figure, the width of the $x$-axis is scaled to the design length $2l$, and the red lines indicate the ROI width.
Fig. 10.
Fig. 10. Confocal detection of a uniform fluorescent plane. The top row (a, b, c) shows the signal as a function of the sample position in air, for the 9 illumination lines. The bottom row (d, e, f) shows the corresponding axial FWHM. The focal lines are labeled with their diffraction order number $m$ and with aranging from blue for $m = - 4$ to yellow for $m = 4.$ (a, b) The measured signal, while a single row on the sensor is used as a confocal detection. (c, d), The measured response for a four-row confocal detection. (e, f) The simulated response for a four-row confocal detection.
Fig. 11.
Fig. 11. Assessment of illumination PSF in a thick fluorescent material. (a) Through focus stack averaged along the $x$ direction. The red crosses mark the focal points. Thebars below the images indicate the intensity value mapping that is adjusted to cover the full dynamic range in each image. (b) The focal points plotted in the $(y,\;z)$ plane (blue circles) including a linear fit through the focal points (blue) which overlaps the object plane (red dashed curve).
Fig. 12.
Fig. 12. A 400 Mpixel maximum intensity projection of a 9 layer multi-focal scan of a tissue micro array (TMA) section labeled using FISH. Every figure zooms in a factor of five, such that the images are shown over a factor of 125 different length scales. The scan is made using an exposure time of 1 ms and four TDI lines. A gamma of 0.7 is used for an enhanced dynamic range. This image is also shown over 100 length scales in Visualization 1. (a) Whole slide scan. The white boxes indicate two regions that displayed in more detail. (b, d) 5$\times$ enlarged. (c, e) 25$\times$ enlarged. (f) 125$\times$ enlarged.
Fig. 13.
Fig. 13. 3D visualizations of a 9 layer multi-focal scan of a tissue micro array (TMA) section labeled using FISH. The scan is made using an exposure time of 10 ms and a single line detector. (a) A frame from a 3D visualization using the Icy software [37] (see Visualization 2). The white line grid has a 10 µm spacing. (b) A maximum intensity projection, where pixels have a assigned according to the layer of maximum intensity, ranging from −5.2 µm (blue) to 5.2 µm (yellow). The FISH probes can be observed to have different positions in $z$. The scale bar represents 10 µm.
Fig. 14.
Fig. 14. A single lane, 9 layer multi-focal scan of an immunofluorescently labeled human rectum tissue section. The scan is made using an exposure time of 10 ms and a single line digital spatial filter. (a) Central layer of the scan, showing the full width of a scan lane. The white box indicates the area that is showed in detail. (b) Detailed view. The white box marks the area displayed in (c-k), which provides a through focus stack of all layers, demonstrating the optical sectioning capability. Although this is a thick tissue section, a high contrast is observed in the central layers.
Fig. 15.
Fig. 15. Determination of the peak signal to noise ratio (pSNR), the peak signal to background ratio (pSBR), and the lateral and axial resolution expressed using the full width at half maximum (FWHM). (a) Maximum intensity projection of the image data of a scan of the FISH slide made using 1 ms exposure time and a four-line TDI acquisition. The scale bar indicates 100 µm. A gamma correction of 0.7 is applied. Three locations are highlighted. (b-d) A zoomed in view to the highlighted areas. The layer with best focus is displayed. The spots location are given including their peak intensity in photo electrons (e). The scale bar indicates 2 µm. The yellow ellipses have a horizontal and vertical diameter equal to the FWHM along the and direction respectively. (e) The image data was classified as background (black) and non-background values (green) or part of a spot. In total, 109 spots where segmented and their location is indicated with a yellow dot. (f-j) Histograms of the pSNR, pSBR and the FWHM along the $x$, and $y$ direction. The red line indicates the mean and the red dashed lines the 5th and 95th percentile respectively. In (h-j) a black line is included indicating the FWHM of the imaging PSF.

Equations (37)

Equations on this page are rendered with MathJax. Learn more.

W ( ρ ) = f ( mod [ K ( ρ ) λ ] ) ,
C m = 0 1 d t exp ( 2 π i m t ) exp ( 2 π i f ( t ) λ ) ,
W m ( ρ ) = m K ( ρ ) ,
K ( ρ ) = Δ y NA ρ y λ + Δ z n 2 NA 2 | ρ | 2 λ ,
f ( t ) = n = 1 N a n cos ( 2 π n t ) ,
Z ( a 0 , , a N ) = m = M M ( | C m | 2 η m ) 2 ,
W ( ρ ) = W y z ( ρ x , ρ y ) + W x ( ρ x ) ,
sin α = 1 R d W x d ρ x .
x = F sin α = 1 NA d W x d ρ x ,
d x = d ( F sin α ) d ρ x d ρ x = 1 NA d 2 W x d ρ x 2 d ρ x .
I f ( x , y ) = ( 1 NA d 2 W x d ρ x 2 ) 1 | U f ( x , y ) | 2 ,
U f ( x , y ) = NA λ 1 ρ x 2 1 ρ x 2 d ρ y A ( ρ x , ρ y ) exp ( 2 π i ρ y y NA λ ) ,
U f ( x , y ) = 2 A NA λ 1 ρ x 2 sinc ( 2 π 1 ρ x 2 y NA λ ) ,
I f ( x , y ) = ( 1 NA d 2 W x d ρ x 2 ) 1 4 A 2 NA 2 λ 2 ( 1 ρ x 2 ) [ sinc ( 2 π 1 ρ x 2 y NA λ ) ] 2 .
I f ( x , y ) = 4 A 2 NA 3 a λ 2 ( 1 [ NA x a ] 2 ) sinc ( 2 π 1 [ NA x a ] 2 y NA λ ) 2 .
d 2 W x d ρ x 2 1 ρ x 2 .
W x ( ρ x ) = 3 4 w NA [ ρ x 2 1 6 ρ x 4 ] ,
x = 3 2 l [ ρ x 1 3 ρ x 3 ] .
ρ x = 2 cos ( π 3 + 1 3 arccos ( x l ) ) .
d 2 W x d ρ x 2 d y | U f ( x , y ) | 2 1 ρ x 2 ,
W x ( ρ x ) = w 2 NA π [ ρ x arcsin ( ρ x ) 1 3 ( 1 ρ x 2 ) 3 2 + 1 ρ x 2 ]
H i l l ( x , y , z ) = m H e x ( y m Δ y , z m Δ z ) ,
I ( M x , M y , y s ) = d x d y d z H e m ( x x , y y , z + Δ z Δ y y ) F ( x , y + y s , z ) H i l l ( x , y , z ) ,
I ( M x , M y , y s ) = m I m ( x , y m Δ y , y s ) ,
I m ( x , y , y s ) = d x d y d z H e m ( x x , y y + y s , z + Δ z Δ y y ) H e x ( y y s , z ) × F ( x , y + m Δ y , z + m Δ z ) ,
I m ( x , y , y s ) = d x d y d z H e m ( x x , y y + y s , z ) H e x ( y y s , z ) × F ( x , y + m Δ y , z + m Δ z ) .
J m ( x i , y i ) = d y I m ( x i , y i , y i + y ) = d x d y d z H ( x i x , y i y , z ) F ( x , y + m Δ y , z + m Δ z ) ,
H ( x , y , z ) = H e m ( x , y , z ) [ d y H e x ( y + y , z ) ]
B ( z ) = d x d y H ( x , y , z ) ,
I m ( x , y , z ) = d x d y H e m ( x x , y y , m Δ z z ) H e x ( y , z m Δ z ) ,
T m ( y , z ) = d x d y d z H e m ( x x , y , z z ) H e x ( y , z ) ,
FOV i l l = 2 F 0 sin η = N 2 λ NA i l l .
z c = s l 1 tan α i l l tan β tan α i l l + tan α d e t .
R ( ϵ 1 , ϵ 2 ) = sin ( 1 2 [ ϕ ( ϵ 1 ) ϕ ( ϵ 2 ) ] ) cos ( 1 2 [ ϕ ( ϵ 1 ) ϕ ( ϵ 2 ) ] ) + C ,
ρ = [ cos θ sin θ sin θ cos θ ] [ M x 0 0 M y ] ( σ σ 0 ) ,
W = g ( mod [ 1 λ A σ x R s ] ) + g ( mod [ 1 λ A σ y R s ] ) + 1 2 B | σ | 2 R S 2 ,
x k l m = F 0 R s [ cos θ sin θ sin θ cos θ ] [ 1 / M x 0 0 1 / M y ] ( A [ k l ] + m Δ B σ 0 R s ) + x 0 ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.