Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Optical modelling of an accommodative light field display system and prediction of human eye responses

Open Access Open Access

Abstract

The spatio-angular resolution of a light field (LF) display is a crucial factor for delivering adequate spatial image quality and eliciting an accommodation response. Previous studies have modelled retinal image formation with an LF display and evaluated whether accommodation would be evoked correctly. The models were mostly based on ray-tracing and a schematic eye model, which pose computational complexity and inaccurately represent the human eye population’s behaviour. We propose an efficient wave-optics-based framework to model the human eye and a general LF display. With the model, we simulated the retinal point spread function (PSF) of a point rendered by an LF display at various depths to characterise the retinal image quality. Additionally, accommodation responses to the rendered point were estimated by computing the visual Strehl ratio based on the optical transfer function (VSOTF) from the PSFs. We assumed an ideal LF display that had an infinite spatial resolution and was free from optical aberrations in the simulation. We tested points rendered at 0–4 dioptres of depths having angular resolutions of up to 4x4 viewpoints within a pupil. The simulation predicted small and constant accommodation errors, which contradict the findings of previous studies. An evaluation of the optical resolution on the retina suggested a trade-off between the maximum achievable resolution and the depth range of a rendered point where in-focus resolution is kept high. The proposed framework can be used to evaluate the upper bound of the optical performance of an LF display for realistically aberrated eyes, which may help to find an optimal spatio-angular resolution required to render a high quality 3D scene.

Published by Optica Publishing Group under the terms of the Creative Commons Attribution 4.0 License. Further distribution of this work must maintain attribution to the author(s) and the published article's title, journal citation, and DOI.

1. Introduction

Accommodation is the mechanism of adjusting the refractive power of the crystalline lens in the eye. A blur pattern in a retinal image drives accommodation so that the retinal image is ‘sharp’ or ‘best focused’ about the fixated object. Accommodation is neurally coupled to binocular vergence, which is a function that rotates the two eyes in opposite directions simultaneously to fuse a visual target at a particular depth into a single image; therefore, an accommodation response is evoked even by a stimulus that elicits only a vergence response and vice versa [13]. In a natural visual environment, this neural link supports quick and robust responses to visual targets so that a viewer can maintain a stable perception of 3D scenes. When observing a stereoscopic image rendered by a conventional 3D display, however, the two signals that drive accommodation may conflict. This is due to the blur pattern on the retina making the eye always focus at the optical distance to the visual target, including a conventional 3D display’s surface, while vergence may widely vary depending on the rendered image depths and where the viewer fixates. This is known as the vergence-accommodation conflict, which causes visual discomfort and fatigue and may also hinder visual performance in some tasks [4,5]. Therefore, developing 3D displays that can provide focus information has become an important research topic.

Light field (LF) displays modulate the position and direction of light rays, aiming to support all visual depth cues including binocular disparity, motion parallax, and focus information. This is attempted mainly by rendering a dense set of directional rays with a sufficiently high spatial and angular resolution. Figure 1 illustrates the basic mechanism of providing focus information by an LF display. When the viewer observes a conventional display (Fig. 1(A)), a focused and sharp retinal image is obtained only when the eye focuses on the display surface; otherwise the retinal image is blurred due to defocus. In other words, the focus information provided by the conventional display drives accommodation only at the display depth. An LF display, on the other hand, may provide more accurate focus information by rendering images with multiple rays that hit the eye’s pupil, as shown in Fig. 1(B). The rays’ directions are controlled so that the rays intersect at the rendered point, imitating natural rays that would emanate from the point. When the eye focuses on the display surface, the individual images formed on the retina by the rays are optically focused; however, their superposition results in a separated or duplicated image that is expected to be recognised as a defocus blur by the visual system. On the other hand, when the eye focuses on the rendered point, the individual images by the rays are optically defocused but overlap perfectly. If the visual system recognises the latter as more ‘focused’ than the former, the correct focus information about the rendered point is provided and accommodation should be driven to the rendered depth. However, the number of rays needed to drive accommodation correctly is not self-evident.

 figure: Fig. 1.

Fig. 1. Basic mechanism of reproducing focus information by LF display. A) Conventional display and retinal images in a schematised eye. The retinal image is characterised by the point spread functions (PSFs) on the right of each diagram. Simulated retinal images of a letter are shown to the right of the corresponding PSFs. In the upper diagram, the eye accommodates at the display surface, whereas the eye accommodates closer than the display surface in the lower diagram. B) An LF display that renders a point at a depth closer than the display with four rays in a pupil (only two are shown), PSFs, and simulated retinal images. In the upper diagram, the eye accommodates at the display surface. In the lower diagram, the eye accommodates at the rendered depth.

Download Full Size | PDF

The idea of making accommodative LF displays was first proposed by Kajiki et al. in 1995 and validated by ray-tracing a simple optical model [6]. Since then, several researchers have implemented the idea in various prototype displays and evaluated their performance in experiments, either by capturing a rendered image with a camera or measuring actual human viewers’ accommodation responses [712]. All these studies reported that accommodation responses were elicited by observing the rendered images, but the responses were not as close to the rendered depths as the responses to real objects. One reason for this may have been the low viewpoint density these prototypes achieved, which compromised three horizontally aligned viewpoints in the pupil at most.

Alternatively to works experimenting with display prototypes, numerous theoretical studies have aimed at developing the analytical evaluation of factors affecting the optical performance of accommodative LF displays. Huang and Hua proposed a framework to model an LF display and a human eye to simulate retinal images [13]. They analytically described the formation of the retinal image and simulated it by ray-tracing a geometrical optical model that comprised the display part and the eye part. Optical resolution on the retina and the predicted accommodation responses were then evaluated by comparing simulated PSFs and modulation transfer functions (MTFs). They predicted accommodation responses considering an integral imaging display that creates 4–16 viewpoints in the pupil of a 3-mm diameter. The predicted responses appeared shifted towards the central depth plane (CDP), which is the optical conjugate of the light source plane, by approximately 0.1–0.2 dioptres. Higher viewpoint densities resulted in smaller predicted shifts. The model was improved further by taking into account the positional sampling and finite pixel resolution in the CDP [14], by generalising it for computational multi-layer LF displays [15], and by unifying the framework to serve both integral imaging displays and computational multi-layer displays [16].

Qin et al. proposed a model aimed at simulating retinal images rendered by a microlens-array-based near-eye LF display, taking into account off-axis ray propagation and using a rigorous calculation of diffraction [17]. Retinal images were simulated by ray-tracing a geometrical optical model that comprised pixels, microlenses, and the Arizona eye model; they later reported predicted accommodation responses to images rendered by that type of LF display [18,19]. Their model had a fixed viewpoint density of 45 viewpoints over the 4-mm diameter pupil and varying CDP depth. The predicted accommodation responses did not generally match the rendered image depths, and significant shifts of the predicted accommodation responses toward the CDP depth were observed.

The methods reviewed above pose certain problems and issues that prevent their direct and systematic use for characterising LF displays and predicting the accommodation response to rendered images. First, these methods are based on ray-tracing, which is computationally inefficient and not suitable for generalisations. Second, the Arizona eye model has been commonly used as the required geometrical optical model of the human eye. This single model replicates chromatic effects and the ‘average’ aberration of human eyes, but it does not represent the peculiarities of the population of human eyes. A handy customised eye model for this purpose is not available [20]. Third, previous studies have not properly modelled and evaluated the accommodation responses for the case of polychromatic light, simulating instead either monochromatic retinal images or coarse spectral sampling. However, polychromatic effects are potentially important, since the human eye suffers from longitudinal chromatic aberration (LCA) greatly and the visual system makes use of it to drive accommodation [21,22].

The prediction of accommodation responses in previous studies may have also been less accurate because of the prediction measures used. These have been based either on the Strehl ratio or on values picked from through-focus MTFs at specific spatial frequencies; the latter does not give a single prediction of the accommodation error but different accommodation errors at different frequencies. On the other hand, the Strehl ratio, while a useful metric of optical performance for slightly aberrated optics, has demonstrated relatively poor performance as an accommodation response predictor [23].

The primary purpose of the current study was to develop a novel framework to model an LF display and a human eye that does not require ray-tracing and can represent the population of aberrated human eyes. Secondarily, the study was aimed at simulating the retinal images of points rendered by an LF display so that (1) the accommodation responses to the rendered points would be predicted and (2) the optical resolution of the focused retinal images would be evaluated.

2. Modelling and evaluation of LF displays

LF displays are essentially aimed at reproducing the spatio-angular intensity distribution of light in a 3D scene. An LF propagating in the half space is usually represented by a four-dimensional function $L(s,t,u,v)$, where two parallel $(s,t)$ and $(u,v)$ planes parametrise the intensity variations along the spatial and angular dimensions of the LF, respectively [24]. Different LF display technologies set these two planes differently. In multi-perspective projection-based LF displays, the two planes are associated with the plane where the projectors are located and the screen plane where the rays recombine [25]. In integral imaging and super multiview displays, a light source plane and a direction-dependent light modulation plane form a pair of planes. In the current study, a microlens-array-based integral imaging display was modelled as a representative LF display technique in line with other related studies, which have considered the same display principle [13,14,16,17]. Other LF display techniques can also be investigated by matching their optical structure and light modulation principle with the corresponding LF parametrisation [13,16].

2.1 Retinal image formation

Figure 2 illustrates the retinal image formation process for an integral-imaging-based LF display. The process begins in the rendering plane which represents a physical source of light. For multiview and integral imaging displays, the rendering plane corresponds to the LCD or OLED screen behind the microlenses. The modulation plane in front of the rendering plane controls the directions of the emitted light with an optical structure such as a lenslet or a parallax barrier. Light emanating from a point source in the rendering plane forms a beam after being spatially limited and typically refracted in the modulation plane. The central depth plane (CDP) is the optical conjugate of the rendering plane that is imaged by the modulation plane. Light rays that emanate from the rendering plane and are modulated on the modulation plane intersect in a plane that includes the rendered 3D image to render a 3D image point, imitating ideal rays that would emanate from that point. Therefore observers are expected to accommodate at this point (see Fig. 1). Each beam of light from a point source in the rendering plane then reaches the eye pupil plane, which is a hypothetical plane that models the optical modulation in the eye, and it forms a subaperture. The beam may be cropped by the pupil and is refracted, forming the image of the point source on the retina, namely the elemental PSF. The final retinal image of the rendered point is the incoherent superposition of the elemental PSFs formed by all rays that enter the eye through the pupil. The number of rays that hit the pupil, which is identically the number of subapertures, is the parameter of the greatest interest in evaluating an accommodative LF display.

 figure: Fig. 2.

Fig. 2. A) An overview of the optical model of an LF display and a human eye. The model consists of six planes: the rendering plane, the modulation plane, the CDP (central depth plane), the depth plane of a rendered image, the eye pupil plane, and the retina. Two rays that render a point image to the right of the CDP are shown. The circles in the eye pupil plane indicate the subapertures formed by the two rays. B) Example of subaperture arrangement for simulation. This example is for a viewpoint density of $3{\times }3$ viewpoints in the pupil. $d_p$ and $d_s$ refer to the diameters of the pupil and the subapertures, and ${\Delta }p$ and ${\Delta }q$ are the intervals of the subapertures in the $p$– and $q$–directions.

Download Full Size | PDF

The position, shape, and size of the subapertures are determined by the setup of the rendering plane and the modulation plane. Let $z_r$ and $z_m$ be the depth of a rendered point and the depth of the modulation plane, respectively (Fig. 2(A)). The positional interval of subapertures $[\Delta {}p,\Delta {}q]$ in Fig. 2(B) is geometrically linked to the sampling interval in the modulation plane $[\Delta {}u,\Delta {}v]$ by

$$\begin{bmatrix} {\Delta}p\\ {\Delta}q \end{bmatrix}=\bigg|\frac{z_r}{z_r-z_m}\bigg| \begin{bmatrix} {\Delta}u\\ {\Delta}v \end{bmatrix},$$
assuming infinitely high resolution on the rendering plane ($(s,t)$ plane), which is equivalent to imposing no positional constraint in determining the locations of the beam emitters. The shape and size of each subaperture is dependent on the modulation structure of the modulation plane. Assuming the origin of the pupil plane is at the centre of the pupil, the amplitude function on the $n$-th subaperture $A_\mathrm {elem}(p,q;n)$ is derived as
$$A_\mathrm{elem}(p,q;n) = \Big|\,\mathfrak{Fr}_{z_m}\big[P_m(u,v)\,\mathfrak{Fr}_{z_0}[\delta(s-s_n,t-t_n)]\big]\Big|,$$
where $\mathfrak {Fr}_z[U(\xi,\eta )]$ denotes the Fresnel diffraction integral [26] as a convolution between an input field $U(\xi,\eta )$ and the convolution kernel for propagation distance $z$, $P_m(u,v)$ represents the modulation function of the modulation structure for the corresponding beam, $\delta (s-s_n,t-t_n)$ denotes the Dirac delta function at $[s_n,t_n]$ modelling the point source in the rendering plane, and $z_0$ and $z_m$ refer to the distance from the rendering plane to the modulation plane and the distance from the modulation plane to the eye pupil plane. The modulation function $P_m(u,v)$ is a simple aperture function or a slit function for displays that modulate light with pinholes or a parallax barrier. For an integral imaging display that controls light’s directions by lenslets, the modulation function includes the phase modulation term corresponding to a microlens, which may also describe its aberration.

In the current study, a hypothetically ideal modulation in the modulation plane was assumed. That is to say, the effects of diffraction and aberration due to the optical modulation structure were neglected, and a point source in the rendering plane was assumed to be imaged in the CDP as an ideal point so that the light sources on the rendering plane can equivalently be treated as ideal point sources in the CDP. This assumption separates the effects of a display’s optical structure from the modelled retinal image formation process. Such a model gives ideal simulation results, which however apply to all LF displays that share the same critical parameters such as viewpoint density and CDP depth, and provide the theoretical maximum of the LF displays’ performance. In our simulations, a subaperture $A_\mathrm {elem}$ was geometrically determined instead of being derived rigorously from Eq. (2) as shown schematically in Fig. 2(A). We first defined three viewpoint densities for the simulation, namely 2$\times$2, 3$\times$3, and 4$\times$4 viewpoints in a 3-mm diameter pupil. To simulate retinal images for a viewpoint density selected from these, the size and positions of the subapertures were determined from the viewpoint density as illustrated in Fig. 2(B). Specifically, the subapertures were set to circular functions that were tangential to each other on a rectangular grid in the eye pupil plane filling the pupil; thus, the diameter of the subapertures $d_s$ was equal to the intervals of the subapertures ${\Delta }p$ and ${\Delta }q$.

Under the assumption of a point source at the CDP and the paraxial approximation, an elemental PSF due to a given beam is obtained by Fourier-transforming the corresponding generalised pupil function, which is nonzero within the corresponding subaperture, as illustrated in Fig. 3(A) [26]. Specifically, the elemental PSF for the $n$-th subaperture at a wavelength $\lambda$ for an eye with a nominal accommodation distance $z_\mathrm {acc}$, namely $\mathrm {ePSF}(x,y;\lambda,n,z_\mathrm {acc})$, is described as

$$\mathrm{ePSF}(x,y;\lambda,n,z_\mathrm{acc}) \propto \Big|\left.\mathcal{F}\!\left[P_\mathrm{sub}(p,q;\lambda,n,z_\mathrm{acc})\right]\right|_{\big(\frac{x}{\lambda{}z_\mathrm{eye}},\frac{y}{\lambda{}z_\mathrm{eye}}\big)}\Big|^2$$
where $\mathcal {F}[\cdot ]$ denotes the Fourier transformation, $P_\mathrm {sub}(p,q;\lambda,n,z_\mathrm {acc})$ is the generalised pupil function of the $n$-th subaperture defined over the whole eye pupil, and $z_\mathrm {eye}$ is the optical distance from the eye pupil plane and the retina. The generalised pupil function is defined as
$$P_\mathrm{sub}(p,q;\lambda,n,z_\mathrm{acc}) = A_\mathrm{sub}(p,q;\lambda,n)\exp{}\Big[\frac{i2\pi{}}{\lambda}\mathit{W}(p,q;\lambda,z_\mathrm{acc})\Big],$$
where $A_\mathrm {sub}(p,q;\lambda,n)$ is the aperture function that describes the amplitude modulation and $\mathit {W}(p,q;\lambda,z_\mathrm {acc})$ is the wavefront aberration function that represents the phase modulation. The generalised pupil function, the aperture function, and the wavefront aberration function are all defined not only in the subaperture but in the whole eye pupil plane.

 figure: Fig. 3.

Fig. 3. Calculation and lateral shift of an elemental PSF. The diagrams show only the $x$ and $p$ dimension for simplicity. The red circle in the lens denotes the $n$-th subaperture and $p_n$ is the position of its centre. A) The geometry of a beam that emanates from the on-axis point in the CDP. The elemental PSF ($\mathrm {ePSF}(x,y;n,z_\mathrm {acc})$) is the retinal image of that point under a nominal accommodation distance of $z_\mathrm {acc}$. $A_\mathrm {elem}$ and $A_\mathrm {sub}$ indicate the amplitude variation of the incident beam and the subsequent amplitude modulation function for calculating the elemental PSF, respectively. B) The shifted elemental PSF when the LF display renders a point at the depth $z_r$. The shift $\Delta {}x_n$ is geometrically determined by $z_\mathrm {CDP}$, $z_\mathrm {eye}$, $z_r$, and $p_n$.

Download Full Size | PDF

The aperture function $A_\mathrm {sub}(p,q;\lambda,n)$ incorporates the amplitude variation of the incident light beam within the corresponding subaperture $A_\mathrm {elem}(p,q;n)$ and a Gaussian apodisation filter in the whole eye pupil plane $A_\mathrm {SCE}(p,q;\lambda ) =10^{-\rho _{\lambda }(p^2+q^2)}$, which models the effect that light passing near the centre of the pupil stimulates cones much more strongly than light passing near the edge of the pupil, referred to as the Stiles–Crawford effect [27,28]. The peak of the effectiveness of the entering light is assumed to be at the centre of the pupil (namely the origin in the $p,q$-plane), and the wavelength-dependent parameter $\rho _{\lambda }$ denotes the peakedness of the effect [29]. The resulting amplitude modulation can then be written as

$$A_\mathrm{sub}(p,q;\lambda,n) = A_\mathrm{pupil}(p,q)A_\mathrm{SCE}(p,q;\lambda)A_\mathrm{elem}(p,q;n),$$
where $A_\mathrm {pupil}(p,q)$ represents the binary circular pupil function $A_\mathrm {pupil}(p,q)=\mathrm {circ}\Big (\frac {\sqrt {p^2+q^2}}{d_p}\Big )$ for a pupil with diameter $d_p$. The wavefront aberration function $W(p,q;\lambda,z_\mathrm {acc})$ incorporates all possible wavefront aberrations compared to the diffraction-limited imaging condition, including not only the fixed aberrations of the eye but also accommodation-dependent defocus and spherical aberration terms. The composition of the wavefront aberration function is further elaborated in Section 2.1.1.

The polychromatic PSF $\mathrm {ePSF_{poly}}(x,y;n,z_\mathrm {acc})$ for each subaperture is found by superposing the elemental PSFs at different wavelengths as

$$\mathrm{ePSF_{poly}}(x,y;n,z_\mathrm{acc}) = {\int}\mathrm{ePSF}(x,y;\lambda,n,z_\mathrm{acc})V(\lambda)\,d\lambda.$$
The monochromatic PSFs are weighted by a luminosity function $V(\lambda )$ to reflect the spectral visual effectiveness [30]. In the current study, the CIE physiologically relevant 2-degree luminosity function for photopic vision was used as $V(\lambda )$ [31], principally because only foveal vision was under consideration. An elemental PSF calculated in this way represents the PSF of the on-axis point source (see Fig. 3(A)). To obtain the retinal PSF of a rendered point, the elemental PSFs are laterally shifted on the retina reflecting the rendered image depth $z_r$, the CDP depth $z_\mathrm {CDP}$, and $z_\mathrm {eye}$ (see Fig. 3(B)). The amount of shift ${\Delta }\textbf {s}_n$ is determined as
$${\Delta}\textbf{s}_n = [\Delta{}x_{n},\Delta{}y_{n}] = z_\mathrm{eye}\bigg(\frac{1}{z_r}-\frac{1}{z_\mathrm{CDP}}\bigg) [p_n,q_n],$$
where $[p_n,q_n]$ is the position of the $n$-th subaperture’s centre.

Finally, laterally shifted polychromatic elemental PSFs are superposed and the retinal PSF of the rendered point under a nominal accommodation distance $z_\mathrm {acc}$ is obtained as

$$\mathrm{PSF_{poly}}(x,y;z_\mathrm{acc}) = \sum_n\mathrm{ePSF_{poly}}(x-{\Delta}x_{n},y-{\Delta}y_{n};n,z_\mathrm{acc})\text{.}$$

2.1.1 Eye wavefront aberrations

The wavefront aberration function $W(p,q;\lambda,z_\mathrm {acc})$ in Eq. (4) includes statistics-based monochromatic aberrations of an unaccommodated eye, the LCA as a wavelength-dependent defocus, the accommodation-related defocus, and the accommodation-dependent additional spherical aberration.

A wavefront aberration function $W(p,q)$ can be described by a Zernike expansion that is a weighted sum of Zernike polynomials as

$$W(p,q)=\sum_{m,n}c^m_nZ^m_n(p,q) ,$$
where $Z^m_n(p,q)$ is a Zernike polynomial, $c^m_n$ is the corresponding Zernike coefficient, and $m$ and $n$ ($n,m{\in }\mathbb {Z}; n{\geq }|m|{\geq }0$) denote the highest order of the polynomial’s radial component and the azimuthal frequency of the polynomial’s sinusoidal component, respectively [32]. With a Zernike expansion, any wavefront aberration function is represented by a set of the Zernike coefficients.

A wavefront aberration function includes defocus of the optical system, which is zero if the system satisfies the imaging condition. In other words, a non-zero defocus means a signed longitudinal shift of the nominal focal point from the CDP depth. Given the nominal accommodation distance $z_\mathrm {acc}$ is provided in metres, the defocus in dioptres $D_\mathrm {acc}(z_\mathrm {acc})$ is found by the following equation:

$$D_\mathrm{acc}(z_\mathrm{acc})=\frac{1}{z_\mathrm{acc}}-\frac{1}{z_\mathrm{CDP}},$$
where $z_\mathrm {CDP}$ is the CDP depth in metres. A dioptric defocus $D_\mathrm {acc}(z_\mathrm {acc})$ contributes to the Zernike defocus term $\textstyle Z^0_2(p,q)$ of the wavefront aberration function. The corresponding coefficient is found through the following equation:
$$c_\mathrm{acc}(z_\mathrm{acc})=\frac{{d_p}^2}{16\sqrt{3}}D_\mathrm{acc}(z_\mathrm{acc})\text{\ ,}$$
where $d_p$ is pupil diameter in millimetres and $S$ is defocus in dioptres [33,34]. Note that $c_\mathrm {acc}(z_\mathrm {acc})$ is in micrometres.

Accommodation primarily changes the shape of the crystalline lens, resulting in an increase in its refractive power and also a systematic decrease in the primary spherical aberration while the other types of aberration statistically show little change with accommodation [22,3537]. Cheng et al. measured changes of aberrations with accommodation and reported that the change in spherical aberration was proportional to the change in accommodation defined by the Zernike defocus [36]. Specifically, the reported slope of the change in the Zernike coefficient for primary spherical aberration $c^0_4$ was $-0.0435$ um/D for 5-mm pupils. The coefficient $c_\mathrm {accSA}$ reflects the contribution of such accommodation-dependent factor to the spherical aberration term $Z^0_4(p,q)$. For a pupil of 5-mm diameter, it is given as

$$c_\mathrm{accSA} = \frac{-0.0435}{z_\mathrm{acc}}\text{.}$$

Human eyes are considerably aberrated, and it greatly affects the quality of the retinal image. Therefore, modelling the human eye with simple ideal optics must over-estimate the retinal resolution. Furthermore, idealising the human eye in the model may also cause an incorrect prediction of the accommodation responses. This is because the empirical accommodation response as a function of the stimulus’s optical distance is widely known to be S-shaped, while the response function of the idealised (diffraction-limited) eye model should be simply the identical line [22,38,39]. This phenomenon, namely the lead and lag of accommodation, is considered to be due to ocular aberrations and pupil constriction [4042]. Similarly, simulating a retinal image for one eye that has the average aberration of the human eye population may also give misleading results. This is because aberrations of the human eye population tend to distribute around zero and thus the hypothetical ‘average’ eye is nearly free of aberrations [43]. In other words, the ‘average’ eye does not represent the human eye population.

To address this issue, we generated multiple instances of an aberrated eye that follow the statistics of measured human eye aberration. Simulated retinal images in these instances, which truly represent the human eye population, were evaluated. To generate corresponding sets of aberration coefficients, we used the statistical model of the aberration of the healthy human eye population by Thibos et al. [44]. In the model, the coefficients for individual eyes are represented as multivariate Gaussian variables with the measured mean and variance of each coefficient and the covariances between all possible pairs of the coefficients. In this study, coefficients representing ten virtual aberrated eyes were generated from the given vector of the mean values of the coefficients and the variance-covariance matrix [45]. Figure 4 shows the resulting Zernike coefficients of the generated virtual eyes with a 3-mm pupil and the mean and $\pm {}2$ standard deviation of each coefficient described in the statistical model. The mean coefficient values are close to zero for most of the Zernike modes, while the coefficient values are much larger for the generated eyes in absolute value.

 figure: Fig. 4.

Fig. 4. Zernike coefficient values for modes 1–13 in OSA/ANSI order. Circles and error bars indicate mean and $\pm {}2$ standard deviation of each coefficient value in the statistical model. Colored lines represent generated Zernike coefficient sets for each virtual eye with a 3-mm pupil.

Download Full Size | PDF

The chromatic aberrations, especially the LCA, also significantly affects the retinal image. The LCA causes the wavelength-dependent defocus of light — specifically, shorter wavelengths of light images at points closer to the lens and vice versa. The defocus due to the LCA is always present in natural viewing, and the visual system utilises it to drive accommodation [21,22]. In contrast to monochromatic aberrations in the human eye population, the profile of the LCA is mostly common across individuals [46,47]. The LCA is measured in dioptres and modelled at a wavelength $\lambda$ by

$$D(\lambda) = A\Bigg(\frac{1}{\lambda-c}-\frac{1}{\lambda_\mathrm{ref}-c}\Bigg),$$
where $A = 633.26$ and $c = 214.10$ [46]. The parameter $\lambda _\mathrm {ref}$ is a reference wavelength in nanometres, at which the retinal image is assumed to be nominally focused. The contribution of the LCA to the defocus term $\textstyle Z^0_2(p,q)$ can be found by converting $D(\lambda )$ to micrometres using Eq. (11).

The target of the simulation was retinal images in eyes with 3-mm pupils, but the statistical data for the accommodation-dependent spherical aberration is provided only for 5-mm pupils. To handle with that, we first generated sets of aberration coefficients based on the statistical data about 4.5-mm pupils. Next, we calculated the accommodation-dependent spherical aberration term and added it to the sets along with the LCA term. Finally, we expanded the coefficients to obtain aberration maps for each eye and cropped them with a 3-mm circular window.

2.2 Prediction of accommodation response

The metrics for the prediction of accommodation responses incorporate not only the retinal image itself but also the subsequent neural factors, reflecting the fact that a retinal defocus blur is detected, processed, and analysed in the visual processing pathway from the retina to the cortex to decide the accommodation response. The visual Strehl ratio is an objective metric that involves the neural transfer function and is known to be one of the best reliable predictors of subjective focus and visual acuity [23,33,4851]. While the Strehl ratio is defined as the ratio of a PSF’s maximum to that of the diffraction-limited PSF, the visual Strehl ratio is computed in the frequency domain (namely the optical transfer function or the modulation transfer function), where it is weighted by a neural factor. In our setting, the visual Strehl ratio computed from the optical transfer function (VSOTF) is given by

$$\mathrm{VSOTF} = \frac{\int_{-\infty}^{\infty}\int_{-\infty}^{\infty}\mathrm{NCSF}(f_x,f_y)\mathrm{OTF_{poly}}(f_x,f_y)df_xdf_y}{\int_{-\infty}^{\infty}\int_{-\infty}^{\infty}\mathrm{NCSF}(f_x,f_y)\mathrm{OTF_{DL}}(f_x,f_y)df_xdf_y},$$
where $\mathrm {OTF_{poly}}$ is the given OTF, $\mathrm {OTF_{DL}}$ is the OTF of the diffraction-limited optics, and $\mathrm {NCSF}$ is the neural contrast sensitivity function [33]. Specifically,
$$\mathrm{OTF_{poly}}(f_x,f_y;z_\mathrm{acc})=\mathcal{F}\big[\mathrm{PSF_{poly}}(x,y;z_\mathrm{acc})\big],$$
and
$$\mathrm{OTF_{DL}}(f_x,f_y) = \mathcal{F}\big[\mathrm{PSF_{polyDL}}(x,y)\big],$$
where $\mathrm {PSF_{polyDL}}(x,y)$ is defined as the special retinal PSF for the eye whose monochromatic aberration coefficients are all zero and is in-focus at the CDP ($z_\mathrm {acc} = z_\mathrm {CDP}$) while the accommodation-related spherical aberration and the wavelength-dependent defocus due to the LCA are present for the sake of consistency in the calculation of polychromatic PSFs.

The NCSF is conceptually the neuro-retinal portion of the contrast sensitivity function (CSF), assuming that the CSF is the product of the eye’s MTF and the NCSF as shown in Fig. 5(B) [52]. The NCSF is obtained either by measuring the contrast sensitivity to interference fringes that are directly formed on the retina [53] or by dividing the CSF by the MTF [52]. We used the NCSF model derived by Watson and Ahumada as shown in Fig. 5(A) [54,55]. The NCSF model incorporates the oblique effect, which is the phenomenon that a typical human observer is less sensitive to the oblique gratings than to the horizontal or vertical gratings.

 figure: Fig. 5.

Fig. 5. A) The NCSF model by Watson & Ahumada (2005, 2012). The model is a two-dimensional real function that describes the neural portion of the contrast sensitivity function (CSF) as a filter. B) A cross section of the NCSF model, MTF, and CSF at vertical frequency = 0. Only half of one row is shown due to symmetry. The MTF (blue line) is calculated for the diffraction-limited eye with a 3-mm pupil. The CSF (black line) is a product of the MTF and the NCSF.

Download Full Size | PDF

Accommodation response to an LF image for each combination of the rendered depths and the viewpoint densities was predicted by through-focus analysis of VSOTF. Figure 6 shows an example of the analysis. In this case, an analysis for one open pupil of an aberrated eye is shown, thus the rendered image depth was only the CDP depth. Retinal PSFs were first computed for varying nominal accommodation distances in front of and behind the image depth and the CDP depth. The frequency domain of the PSFs, namely the OTFs, were then computed; finally, VSOTF values as a function of the nominal accommodation distance were obtained. The nominal accommodation distance at which VSOTF takes the peak value is the predicted accommodation response. Therefore, in Fig. 6, an accommodation response is predicted at $+0.2$D relative to the image depth.

 figure: Fig. 6.

Fig. 6. Example of through-focus analysis for one open pupil of an aberrated eye. The CDP depth was 2D. OTFs are visualised in a manner that the brightness and hue correspondingly depict the OTF’s modulus and angle.

Download Full Size | PDF

3. Simulation results

In this section, (1) predictions of accommodation response to rendered points and (2) the optical resolution of their images on the retina are reported for a wide range of depths rendered by an integral-imaging-based LF display with various viewpoint densities. Through-focus analysis was employed to predict the accommodation response, and optical resolutions on the retina at various conditions were compared between each of them at the best focused state found by the through-focus analysis.

The CDP depth of the LF display model was fixed at 2 dioptres. The three viewpoint densities, i.e. 2$\times$2, 3$\times$3, and 4$\times$4 viewpoints within the eye pupil of a 3-mm diameter, were tested. The pupil diameter of 3 mm was chosen as a typical pupil size during binocularly viewing a typically bright display, having a luminance of 100–200 $\mathrm {cd/m^2}$ [1316,56]. The rendered image depth was one of nine depths in front of and behind the CDP, namely $-2.0$D, $-1.5$D, $\dots \text {,}$ $0$D, $\dots \text {,}$ or $+2.0$D relative to the CDP depth. The intervals of the nominal accommodation distances for the through-focus analysis were set to be $0.2$D. The spectrum of visible light was sampled at 400 nm and 10 nm increments up to 700 nm. We set the reference wavelength at 550 nm, at which the luminosity function approximately reaches its peak [57]. The whole simulation was implemented by customising the ISETbio toolbox [58] on MATLAB.

3.1 Prediction of accommodation response

The through-focus analysis was repeated for all combinations of the eyes, viewpoint densities, and rendered image depths.

3.1.1 Natural view

Firstly, for reference, we conducted the simulation for the case of natural view, where no subaperture was set and thus the ‘rendered’ image depth was only the CDP depth. Figure 7 shows the through-focus VSOTFs calculated for the aberrated eyes and the average eye; their peak positions are shown as vertical lines, which indicates the predicted accommodation depths for the eyes. The aberrated eyes had generally different through-focus response profiles from each other (the grey curves), but none of the ten eyes was predicted to focus at a depth apart from the image depth ($=$ the CDP depth) by more than $0.2$D. To see the average response of the aberrated eyes, the average position of the peaks for the aberrated eyes is shown with the vertical red line. The average response was close to $0$D with a slight negative shift in dioptres. It means the eyes were, on average, predicted to focus at a depth close to the CDP when observing a point on the CDP naturally.

 figure: Fig. 7.

Fig. 7. Through-focus VSOTF for natural view. The grey curves indicate metric values as functions of relative accommodation, namely additional defocus to the rendered image depth. The grey vertical lines show their peaks. The red vertical line is the average of the peak positions over the eyes. The blue curve and vertical line represent the metric values and its peak for the average eye. The image depth was that of the CDP (2 dioptres).

Download Full Size | PDF

The through-focus VSOTF for the average eye (the blue curve) was much greater than that for the aberrated eyes around its peak and almost reaching the value of 1, suggesting an eye with the average aberration is very close to the diffraction-limited eye when it is in focus. The predicted accommodation response for the average eye was, however, essentially the same as the average response of the aberrated eyes.

3.1.2 Rendered 3D image

Next, we simulated retinal PSFs of images rendered at a wide range of depths relative to the CDP, and accommodation responses to the rendered image were predicted by the through-focus analysis of VSOTF. The simulation was conducted for three configurations where 2$\times$2, 3$\times$3, or 4$\times$4 viewpoints were in the pupil of 3 mm diameter.

Figure 8 shows the through-focus plots of VSOTF for each image depth and view density. The mean predicted responses of the aberrated eyes were close to the rendered image depth for all viewpoint densities and image depths. In other words, the accommodation error, i.e. the dioptric difference between the image depth and the predicted accommodation distance, was predicted to be small regardless of the rendered image depth and the viewpoint density. The accommodation errors tended to be small but non-zero and constantly negative. This means that the aberrated eyes were predicted to focus at depths slightly farther than the image depth regardless of the rendered depth of the image.

 figure: Fig. 8.

Fig. 8. Through-focus VSOTF for the viewpoint densities of 2$\times$2, 3$\times$3, and 4$\times$4 viewpoints in a 3-mm pupil. Data are plotted in the same manner as Fig. 7. The green vertical lines indicate the CDP depth. Note that some parts of the through-focus ranges are missing in some panels on the top because we assumed the eyes would not accommodate hyperopically, namely its absolute accommodation distance should not be ‘farther than infinity’. Note that the scales of y-axes are not unified for better visibility of through-focus variations in each panel.

Download Full Size | PDF

The tendency of the negative errors may be due to the effect of the spherical aberration in the eye. That is to say, focusing at slightly farther than the nominally in-focus depth may improve the retinal image quality, or increase VSOTF in this case, under presence of the spherical aberration [4042]. The constant accommodation errors across the image depths contradict the previously reported simulation results in which the accommodation error was reported to increase as the depth difference between the rendered image and the CDP grows.

The accommodation responses predicted for the average eye (the blue vertical lines in Fig. 8) were very close to the average of the predicted accommodation responses for the aberrated eyes (the red vertical lines in the same figure). This similarity suggests that a through-focus analysis of VSOTF on the average eye may be sufficient to predict accommodation responses, although the average eye is very close to a diffraction-limited system and thus poorly represents aberrated real eyes.

In summary, the accommodation responses were predicted to be close to the rendered image depths, and the predicted accommodation errors were roughly constant across the image depths. In addition to that, increasing the viewpoint density did not affect the accommodation errors, at least in the configurations we tested. It can straightforwardly be interpreted that even the lowest viewpoint density we tested, namely 2$\times$2 viewpoints in the 3-mm diameter pupil, is fairly enough to elicit an accommodation response as correctly as in the natural view.

These observations generally contradict the findings that have previously been reported [13,16,18,19]. Huang and Hua predicted smaller accommodation errors in configurations with higher viewpoint densities [13], and Qin et al. reported that systematic accommodation errors towards the CDP depth were predicted from through-focus analyses of the Strehl ratio, even though a large number of viewpoints (45 viewpoints over a 4-mm pupil) was assumed in their simulation [18,19]. In the current study, however, the predicted accommodation responses did not show any clear difference between the viewpoint densities, and the predicted accommodation errors were small and constant across the rendered image depths.

There are two possible points that may explain the difference. One is the effect of display diffraction and aberration, which were ignored in the simulation in the current study but included in the simulations in all previous studies. Including them into the model must always make the elemental PSFs more blurred, since a point source at the CDP is an idealised simplification of the image of a point source on the rendering plane. Nevertheless, it does not directly imply that including the effect of display diffraction and aberration into the simulation makes the results closer to these in the previous studies. The other is the prediction methods of accommodation responses. In the current study, VSOTF was used as the predictor metric for accommodation response, while the previous studies used through-focus Strehl ratios and values picked from through-focus MTFs.

We also analysed through-focus Strehl ratios for the viewpoint densities and rendered image depths (Fig. 9). The accommodation responses predicted from through-focus Strehl ratios were essentially identical to those from the through-focus VSOTF except the cases for the viewpoint density of 2${\times }$2 viewpoints in the pupil. That is to say, the through-focus Strehl ratio predicted that the accommodation responses would be close to the image depths rather than the CDP depth in the cases for the viewpoint densities of 3$\times$3 or 4$\times$4 viewpoints in a 3-mm pupil. For the viewpoint density of 2$\times$2 viewpoints in the pupil, however, the results were largely different from the other viewpoint densities. Specifically, the predicted accommodation depths were close to the image depths only if the rendered images were in a depth range of ${\pm }1.0$D relative to the CDP; otherwise, the predicted accommodation responses were close to the CDP rather than the image depths. The tendency of small negative accommodation errors was observed in the through-focus Strehl ratio as well as in the through-focus VSOTF except in the cases where the image depths were apart from the CDP by more than 1.0D for the viewpoint densities of 2$\times$2 viewpoints in the pupil. The predicted accommodation responses on the average eye were very similar to those on the aberrated eyes, as observed also in the through-focus analysis of VSOTF.

 figure: Fig. 9.

Fig. 9. Through-focus Strehl ratio for viewpoint densities of $2{\times }2$, $3{\times }3$, and $4{\times }4$ viewpoints in a 3-mm pupil. Data are shown in the same manner as Fig. 8.

Download Full Size | PDF

Figure 10 shows an example case in which the predictions from the through-focus VSOTF and Strehl ratio were largely different; VSOTF predicted the accommodation response around the image depth, while the Strehl ratio predicted the response at the CDP depth. When the eye accommodated at the image depth, the PSF spread widely and thus the simulated retinal image was blurred. On the other hand, when the CDP was accommodated, the PSF consisted of four separate and clear peaks, hence the retinal image appeared as a superposition of four clear images with positional differences, which eventually looked hardly recognisable. Since only the peak amplitude of a PSF defines the Strehl ratio, its highest value seemed to be obtained in the case when the CDP was accommodated. There is much evidence to support the higher validity of VSOTF over the Strehl ratio as a predictor of accommodation response, but all of these studies, to the best of our knowledge, tested it for eyes with a natural pupil. Hence, the validity of VSOTF is not perfectly supported in predicting the accommodation response from ‘irregular’ PSFs that are rendered by an LF display. However, this also applies to other metrics such as the Strehl ratio. Considering that the Strehl ratio is usually used only for optical systems with little aberration, the Strehl ratio may not be a suitable metric to predict the accommodation response from PSFs rendered by an LF display.

 figure: Fig. 10.

Fig. 10. Through-focus VSOTF and Strehl ratio for an image rendered at the CDP depth $-2.0$D and PSFs and simulated retinal images at the two accommodation states in the through-focus range where the eye accommodates at the rendered image and the CDP. The PSFs and retinal images are for one of the aberrated eyes.

Download Full Size | PDF

3.2 Optical resolution of rendered image

The two-dimensional MTFs for the in-focus accommodation states predicted by the VSOTF were averaged across meridians, providing radial MTFs [43,59], so that the MTFs were visualised and compared with each other more simply. Figure 11 shows the obtained radial MTFs, which describe the optical resolution on the retina for each rendered image depth and viewpoint density. The mean MTF for the aberrated eyes and the MTF for the average eye are separately plotted. The MTFs for the average eye were better than the mean MTFs for the aberrated eyes in almost all cases. This indicates that the simulation on the eye model with the average aberration overestimates retinal optical resolution.

 figure: Fig. 11.

Fig. 11. Radial MTFs of the retinal images at predicted accommodation by VSOTF for each image depth and viewpoint density. The black plots with grey shades are the mean and $\pm {}1$ standard deviation of the radial MTFs for the aberrated eyes. The blue plots are the radial MTFs for the average eye.

Download Full Size | PDF

In general, better MTFs were simulated for the cases with the smaller depth difference between the rendered image and the CDP, especially when the viewpoint density was low. This tendency was less clear for the cases with higher viewpoint densities. Specifically, for the cases with the lowest viewpoint density — namely 2$\times$2 viewpoints in the pupil — the optical resolution drastically dropped as the depth difference between the rendered image and the CDP grew, even though the retinal image was focused. It can be interpreted that the depth of field (DOF) of the LF display, or the range of a rendered depth where the resolution of the in-focus retinal image is kept high, is shallow for a configuration with a low viewpoint density. On the other hand, with the high viewpoint density, the drop of the in-focus optical resolution related to the rendered image depth relative to the CDP was less distinctive. In other words, a wider DOF was simulated for a configuration with a high viewpoint density.

To visualise the relation between the viewpoint density and the optical resolution on the retina, we set the reference contrast gain of 0.05 as a threshold and plotted the cut-off frequencies for the three viewpoint densities (Fig. 12). We chose the reference gain of 0.05 because the contrast gain dropped to around that value on average in the aberrated eyes at 60 CPD, which is dogmatically the highest spatial frequency resolvable for human observers (Fig. 12(A)). Figure 12(B) shows the cut-off frequencies for the three viewpoint densities as functions of the image depth relative to the CDP. As already described, a configuration with the low viewpoint density (2$\times$2 viewpoints) was predicted to achieve a high resolution, but the DOF was small. On the other hand, a configuration with a high viewpoint density (4$\times$4 viewpoints) can achieve low resolution at best, but the DOF was greater than the lower viewpoint densities.

 figure: Fig. 12.

Fig. 12. (A) Radial MTF for open 3-mm pupil at the in-focus state predicted by VSOTF. The blue line represents the radial MTF for the average eye. The black line and grey shaded area show the mean and $\pm$1 standard deviation of the radial MTFs for the aberrated eyes. (B) Cut-off frequencies for reference gain of 0.05 for the three viewpoint densities.

Download Full Size | PDF

The best optical resolution, which was achieved when the image was rendered at the CDP depth, was better in the cases with the low viewpoint density. A drop of the maximum resolution with increasing viewpoint density was also reported previously [13], but the drop observed in the current study was not as drastic as in their report.

3.3 Summary

We simulated through-focus PSFs for a wide range of the image depths and the viewpoint densities of 2$\times$2, 3$\times$3, or 4$\times$4 viewpoints in a 3-mm diameter pupil assuming no effect of display diffraction and aberration. The accommodation responses to the rendered images were predicted by the through-focus analysis of VSOTF, and the optical resolution on the retina was also assessed.

The predicted accommodation responses were close to the rendered image depths even for the cases where the image depth was apart from the CDP depth. Importantly, the predicted accommodation errors were constantly small across the viewpoint densities. This means that even the lowest viewpoint density we tested — namely 2$\times$2 viewpoints in the pupil — may be enough to elicit correct accommodation to rendered images at a wide range of depths if the LF display had the ideal optics so that the effects of its diffraction and aberration were negligible.

The analysis of the optical resolution on the retina showed the advantage of a wide DOF for the LF display with a high viewpoint density. In other words, increasing the viewpoint density extended the range of a rendered image depth in which the best retinal image can be kept.

To sum up, under the assumption of the absence or negligible effects of display diffraction and aberration, it was suggested that the viewpoint density may not be needed to be that high to elicit the correct accommodation responses to rendered images; however, it was also inferred that a low viewpoint density may limit the range of a rendered depth in which the in-focus resolution is kept high on the retina.

4. Limitations of the study

The current study is limited principally at several points. The most important limitation is the absence of display diffraction and aberration. As discussed earlier, ignoring these effects and assuming an ideal point source at the CDP must overestimate the optical resolution on the retina. Hence, the simulation results estimate the retinal image resolution only for an LF display with little effects of diffraction and aberration, which seems to be hardly achievable. Simultaneously, it can be inferred that no real LF displays can achieve a retinal image resolution that is better than the results simulated in the proposed way. On the other hand, the effects of display diffraction and aberration on accommodation response cannot be simply guessed from the logic that including them would make the elemental PSFs less sharp, because the visual system always observes their superposition.

Another limitation is the absence of pixels in the model. In the current study, the point source was assumed to be infinitesimal and positional sampling was ignored. However, in a practical LF display, the image resolution is always regulated by the display’s pixel sampling. Pixel sampling strongly affects the image resolution a viewer may observe in combination with the rendered depth [60,61]. Specifically, the pixel sampling determines the depth range in which the best pixel resolution can be obtained for the viewer. A further study is needed to integrate the optical effects of diffraction and aberration in a display and human eye into the discussion of the pixel resolution and the rendered depth, although an initial attempt has been made by Huang and Hua by extending their study [14].

The last important limitation to be pointed out is the paraxial approximation. The calculation of PSFs in the proposed framework gives an accurate simulation only in the paraxial region. In other words, the calculation cannot simply be used to assess the optical image on the peripheral retina, which is formed by off-axis light rays. In addition to that, because the retina is not spatially uniform physiologically and perceptually, the functional role of the non-foveal or peripheral region in accommodation must be different from that of the foveal region [62]. Therefore, involving the functional role of the non-foveal or peripheral retina in accommodation may require not only an appropriate calculation method for the retinal image, but also an evaluation method that correctly reflects the characteristics of the peripheral visual field.

5. Conclusion

In the current study, we proposed a novel simulation framework to model an LF display and an observer’s eye to simulate the retinal image, which is totally free from ray-tracing. The proposed framework has the advantages of (1) being free from ray-tracing, thus computationally efficient; (2) having the capability to include realistic aberration patterns of the human eye population; and (3) ensuring the rigorous modelling of chromatic effects in the visual system.

In addition to that, an optical metric that is known to predict accommodation response well was used to assess whether the accommodation would be elicited correctly. The simulation based on the proposed model showed that accommodation was expected to be elicited fairly close to the rendered depths even in a configuration with a relatively low viewpoint density, e.g. 2$\times$2 views within a 3 mm pupil. However, increasing the viewpoint density seemed to extend the depth range of the rendered image in which in-focus retinal resolution is kept high.

The simulation was based on the assumption of hypothetically ideal display optics. As such, it did not quantify how the aberrations and diffraction on the display side affect accommodation responses and the retinal optical resolution. Further studies are needed to extend the simulation to include these effects assuming a specific setup of an actual LF display ; then predictions from the extended simulation can be compared to empirical measurement results on the real LF display. Such a comparison will critically verify the proposed model and accommodation prediction.

Funding

H2020 Marie Skłodowska-Curie Actions (764951); Academy of Finland (325530).

Disclosures

The authors declare no conflicts of interest.

Data Availability

Data underlying the results presented in this paper are available in Refs. [44,45]. Customised code for computations in this paper is available via our website [63].

References

1. S. R. Bharadwaj, “Accommodation mechanisms,” in Handbook of Visual Optics, vol. 1P. Artal, ed. (CRC Press, 2017), chap. 24, pp. 363–376.

2. T. Inoue and H. Ohzu, “Accommodative responses to stereoscopic three-dimensional display,” Appl. Opt. 36(19), 4509 (1997). [CrossRef]  

3. M. Torii, Y. Okada, K. Ukai, J. S. Wolffsohn, and B. Gilmartin, “Dynamic measurement of accommodative responses while viewing stereoscopic images,” J. Mod. Opt. 55(4-5), 557–567 (2008). [CrossRef]  

4. D. M. Hoffman, A. R. Girshick, K. Akeley, and M. S. Banks, “Vergence–accommodation conflicts hinder visual performance and cause visual fatigue,” J. Vis. 8(3), 33 (2008). [CrossRef]  

5. P. V. Johnson, J. A. Q. Parnell, J. Kim, C. D. Saunter, G. D. Love, and M. S. Banks, “Dynamic lens and monovision 3D displays to improve viewer comfort,” Opt. Express 24(11), 11808 (2016). [CrossRef]  

6. Y. Kajiki, H. Yoshikawa, and T. Honda, “3D display with focused light array,” in Proceedings of SPIE, vol. 2652S. A. Benton, ed. (San Jose, CA, USA, 1996), pp. 106–116.

7. Y. Miyanishi, E. Sahin, and A. Gotchev, “Providing focus cues by light field displays: A review,” European Light Field Imaging Workshop (2019).

8. T. Honda, Y. Kajiki, K. Susami, T. Hamaguchi, T. Endo, T. Hatada, and T. Fujii, “Three-dimensional display technologies satisfying “super multiview condition”,” Proc. SPIE 10298, 102980B (2001). [CrossRef]  

9. T. Sudo, H. Morishima, T. Osaka, and N. Taniguchi, “3D display using intersection of light beams,” in SPIE Electronic Imaging, vol. 3957J. O. Merritt, S. A. Benton, A. J. Woods, and M. T. Bolas, eds. (International Society for Optics and Photonics, San Jose, CA, USA, 2000), p. 215.

10. H. Nate, “(Original title in Japanese) Super multiview display and its evaluation,” Vis. 15, 87–91 (2003).

11. H. Mizushina, J. Nakamura, Y. Takaki, and H. Ando, “Super multi-view 3D displays reduce conflict between accommodative and vergence responses,” J. Soc. Inf. Disp. 24(12), 747–756 (2016). [CrossRef]  

12. H. Hiura, T. Mishina, J. Arai, and Y. Iwadate, “Accommodation response measurements for integral 3D image,” in SPIE Electronic Imaging, vol. 9011A. J. Woods, N. S. Holliman, and G. E. Favalora, eds. (International Society for Optics and Photonics, 2014), p. 90111H.

13. H. Huang and H. Hua, “Systematic characterization and optimization of 3D light field displays,” Opt. Express 25(16), 18508 (2017). [CrossRef]  

14. H. Huang and H. Hua, “Effects of ray position sampling on the visual responses of 3D light field displays,” Opt. Express 27(7), 9343–9360 (2019). [CrossRef]  

15. M. Xu and H. Hua, “Systematic method for modeling and characterizing multilayer light field displays,” Opt. Express 28(2), 1014–1036 (2020). [CrossRef]  

16. M. Xu, H. Huang, and H. Hua, “Analytical model for the perceived retinal image formation of 3D display systems,” Opt. Express 28(25), 38029–38048 (2020). [CrossRef]  

17. Z. Qin, P. Chou, J. Wu, Y. Chen, C. Huang, N. Balram, and Y. Huang, “Image formation modeling and analysis of near-eye light field displays,” J. Soc. Inf. Disp. 27(4), 238–250 (2019). [CrossRef]  

18. Z. Qin, J.-Y. Wu, P.-Y. Chou, Y.-T. Chen, C.-T. Huang, N. Balram, and Y.-P. Huang, “Revelation and addressing of accommodation shifts in microlens array-based 3D near-eye light field displays,” Opt. Lett. 45(1), 228–231 (2020). [CrossRef]  

19. Z. Qin, J.-Y. Wu, P.-Y. Chou, J. Xie, C.-T. Huang, J.-L. Chen, H.-P. D. Shieh, and Y.-P. Huang, “68-1: Investigation on defocusing-induced accommodation shift in microlens array-based near-eye light field displays,” in SID’s Digest of Technical Papers, vol. 51 (Wiley, 2020), pp. 1009–1012.

20. J. Tabernero, “Customized eye models,” in Handbook of Visual Optics, vol. 2P. Artal, ed. (CRC, 2017).

21. S. A. Cholewiak, G. D. Love, and M. S. Banks, “Creating correct blur and its effect on accommodation,” J. Vis. 18(9), 1 (2018). [CrossRef]  

22. A. J. Del Águila-Carrasco, P. B. Kruger, F. Lara, and N. López-Gil, “Aberrations and accommodation,” Clin. Exp. Optom. 103(1), 95–103 (2020). [CrossRef]  

23. X. Cheng, A. Bradley, and L. N. Thibos, “Predicting subjective judgment of best focus with objective image quality metrics,” J. Vis. 4(4), 7 (2004). [CrossRef]  

24. N. Balram and I. Tošić, “Light-field imaging and display systems,” Inf. Disp. 32(4), 6–13 (2016). [CrossRef]  

25. R. Bregovic, E. Sahin, S. Vagharshakyan, and A. Gotchev, “Signal processing methods for light field displays,” in Handbook of Signal Processing Systems, (Springer International Publishing, Cham, 2019), pp. 3–50.

26. J. Goodman, Introduction to Fourier Optics (W. H. Freeman, 2017), 4th ed.

27. W. S. Stiles, B. H. Crawford, and J. H. Parsons, “The luminous efficiency of rays entering the eye pupil at different points,” Proc. R. Soc. Lond. B. 112(778), 428–450 (1933). [CrossRef]  

28. G. Westheimer, “Directional sensitivity of the retina: 75 years of Stiles–Crawford effect,” Proc. R. Soc. London, Ser. B 275(1653), 2777–2786 (2008). [CrossRef]  

29. T. T. Berendschot, J. van de Kraats, and D. van Norren, “Wavelength dependence of the Stiles–Crawford effect explained by perception of backscattered light from the choroid,” J. Opt. Soc. Am. A 18(7), 1445–1451 (2001). [CrossRef]  

30. “ISO/CIE 11664-1:2019 Colorimetry — Part 1: CIE standard colorimetric observers,” (2019).

31. A. Stockman, H. Jägle, M. Pirzer, and L. T. Sharpe, “The dependence of luminous efficiency on chromatic adaptation,” J. Vis. 8(16), 1 (2008). [CrossRef]  

32. L. N. Thibos, R. A. Applegate, J. T. Schwiegerling, and R. Webb, and VSIA Standards Taskforce Members, “Standards for reporting the optical aberrations of eyes,” (2000).

33. L. N. Thibos, X. Hong, A. Bradley, and R. A. Applegate, “Accuracy and precision of objective refraction from wavefront aberrations,” J. Vis. 4(4), 9 (2004). [CrossRef]  

34. “ANSI Z80.28-2017 Ophthalmics - Methods of Reporting Optical Aberrations of Eyes,” (2017).

35. J. C. He, S. A. Burns, and S. Marcos, “Monochromatic aberrations in the accommodated human eye,” Vision Res. 40(1), 41–48 (2000). [CrossRef]  

36. H. Cheng, J. K. Barnett, A. S. Vilupuru, J. D. Marsack, S. Kasthurirangan, R. A. Applegate, and A. Roorda, “A population study on changes in wave aberrations with accomodation,” J. Vis. 4(4), 3 (2004). [CrossRef]  

37. N. López-Gil, V. Fernández-Sánchez, R. Legras, R. Montés-Micó, F. Lara, and J. L. Nguyen-Khoa, “Accommodation-related changes in monochromatic aberrations of the human eye as a function of age,” Invest. Ophthalmol. Vis. Sci. 49(4), 1736–1743 (2008). [CrossRef]  

38. T. Buehren and M. J. Collins, “Accommodation stimulus-response function and retinal image quality,” Vision Res. 46(10), 1633–1645 (2006). [CrossRef]  

39. V. Labhishetty, S. A. Cholewiak, A. Roorda, and M. S. Banks, “Lags and leads of accommodation in humans: Fact or fiction?” J. Vis. 21(3), 21 (2021). [CrossRef]  

40. N. Lopez-Gil and V. Fernandez-Sanchez, “The change of spherical aberration during accommodation and its effect on the accommodation response,” J. Vis. 10(13), 12 (2010). [CrossRef]  

41. L. N. Thibos, A. Bradley, and N. López-Gil, “Modelling the impact of spherical aberration on accommodation,” Ophthalmic & Physiol. Opt. journal Br. Coll. Ophthalmic Opt. 33(4), 482–496 (2013). [CrossRef]  

42. N. López-Gil, J. Martin, T. Liu, A. Bradley, D. Díaz-Mu noz, and L. N. Thibos, “Retinal image quality during accommodation,” Ophthalmic & Physiol. Opt. journal Br. Coll. Ophthalmic Opt. 33(4), 497–507 (2013). [CrossRef]  

43. L. N. Thibos, X. Hong, A. Bradley, and X. Cheng, “Statistical variation of aberration structure and image quality in a normal population of healthy eyes,” J. Opt. Soc. Am. A 19(12), 2329–2348 (2002). [CrossRef]  

44. L. N. Thibos, A. Bradley, and X. Hong, “A statistical model of the aberration structure of normal, well-corrected eyes,” Ophthalmic & Physiol. Opt. journal Br. Coll. Ophthalmic Opt. 22(5), 427–433 (2002). [CrossRef]  

45. L. N. Thibos, “Retinal image quality for virtual eyes generated by a statistical model of ocular wavefront aberrations,” Ophthalmic & Physiol. Opt. journal Br. Coll. Ophthalmic Opt. 29(3), 288–291 (2009). [CrossRef]  

46. L. N. Thibos, M. Ye, X. Zhang, and A. Bradley, “The chromatic eye: a new reduced-eye model of ocular chromatic aberration in humans,” Appl. Opt. 31(19), 3594 (1992). [CrossRef]  

47. M. Nakajima, T. Hiraoka, Y. Hirohara, T. Oshika, and T. Mihashi, “Verification of the lack of correlation between age and longitudinal chromatic aberrations of the human eye from the visible to the infrared,” Biomed. Opt. Express 6(7), 2676–2694 (2015). [CrossRef]  

48. J. D. Marsack, L. N. Thibos, and R. A. Applegate, “Metrics of optical quality derived from wave aberrations predict visual performance,” J. Vis. 4(4), 8 (2004). [CrossRef]  

49. J. Tarrant, A. Roorda, and C. F. Wildsoet, “Determining the accommodative response from wavefront aberrations,” J. Vis. 10(5), 4 (2010). [CrossRef]  

50. J. Martin, B. Vasudevan, N. Himebaugh, A. Bradley, and L. Thibos, “Unbiased estimation of refractive state of aberrated eyes,” Vision Res. 51(17), 1932–1940 (2011). [CrossRef]  

51. A. Ravikumar, E. J. Sarver, and R. A. Applegate, “Change in visual acuity is highly correlated with change in six image quality metrics independent of wavefront error and/or pupil diameter,” J. Vis. 12(10), 11 (2012). [CrossRef]  

52. R. Michael, O. Guevara, M. de la Paz, J. Alvarez de Toledo, and R. I. Barraquer, “Neural contrast sensitivity calculated from measured total contrast sensitivity and modulation transfer function,” Acta Ophthalmol. 89(3), 278–283 (2011). [CrossRef]  

53. F. W. Campbell and D. G. Green, “Optical and retinal factors affecting visual resolution,” The J. Physiol. 181(3), 576–593 (1965). [CrossRef]  

54. A. B. Watson and A. J. Ahumada, “Modeling acuity for optotypes varying in complexity,” J. Vis. 12(10), 19 (2012). [CrossRef]  

55. A. B. Watson and A. J. Ahumada Jr, “A standard model for foveal detection of spatial contrast,” J. Vis. 5(9), 6 (2005). [CrossRef]  

56. A. B. Watson and J. I. Yellott, “A unified formula for light-adapted pupil size,” J. Vis. 12(10), 12 (2012). [CrossRef]  

57. S. Ravikumar, L. N. Thibos, and A. Bradley, “Calculation of retinal image quality for polychromatic light,” J. Opt. Soc. Am. A 25(10), 2395–2407 (2008). [CrossRef]  

58. N. P. Cottaris, H. Jiang, X. Ding, B. A. Wandell, and D. H. Brainard, “A computational-observer model of spatial contrast sensitivity: Effects of wave-front-based optics, cone-mosaic structure, and inference engine,” J. Vis. 19(4), 8 (2019). [CrossRef]  

59. A. B. Watson, “A formula for the mean human optical modulation transfer function as a function of pupil size,” J. Vis. 13(6), 18 (2013). [CrossRef]  

60. H. Hoshino, F. Okano, H. Isono, and I. Yuyama, “Analysis of resolution limitation of integral photography,” J. Opt. Soc. Am. 15(8), 2059 (1998). [CrossRef]  

61. M. Zwicker, A. Vetro, S. Yea, W. Matusik, H. Pfister, and F. Durand, “Resampling, antialiasing, and compression in multiview 3-D displays,” IEEE Signal Process. Mag. 24(6), 88–96 (2007). [CrossRef]  

62. V. Labhishetty, S. A. Cholewiak, and M. S. Banks, “Contributions of foveal and non-foveal retina to the human eye’s focusing response,” J. Vis. 19(10), 121 (2019). [CrossRef]  

63. “3D Media Group,” https://research.tuni.fi/3dmedia/.

Data Availability

Data underlying the results presented in this paper are available in Refs. [44,45]. Customised code for computations in this paper is available via our website [63].

44. L. N. Thibos, A. Bradley, and X. Hong, “A statistical model of the aberration structure of normal, well-corrected eyes,” Ophthalmic & Physiol. Opt. journal Br. Coll. Ophthalmic Opt. 22(5), 427–433 (2002). [CrossRef]  

45. L. N. Thibos, “Retinal image quality for virtual eyes generated by a statistical model of ocular wavefront aberrations,” Ophthalmic & Physiol. Opt. journal Br. Coll. Ophthalmic Opt. 29(3), 288–291 (2009). [CrossRef]  

63. “3D Media Group,” https://research.tuni.fi/3dmedia/.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (12)

Fig. 1.
Fig. 1. Basic mechanism of reproducing focus information by LF display. A) Conventional display and retinal images in a schematised eye. The retinal image is characterised by the point spread functions (PSFs) on the right of each diagram. Simulated retinal images of a letter are shown to the right of the corresponding PSFs. In the upper diagram, the eye accommodates at the display surface, whereas the eye accommodates closer than the display surface in the lower diagram. B) An LF display that renders a point at a depth closer than the display with four rays in a pupil (only two are shown), PSFs, and simulated retinal images. In the upper diagram, the eye accommodates at the display surface. In the lower diagram, the eye accommodates at the rendered depth.
Fig. 2.
Fig. 2. A) An overview of the optical model of an LF display and a human eye. The model consists of six planes: the rendering plane, the modulation plane, the CDP (central depth plane), the depth plane of a rendered image, the eye pupil plane, and the retina. Two rays that render a point image to the right of the CDP are shown. The circles in the eye pupil plane indicate the subapertures formed by the two rays. B) Example of subaperture arrangement for simulation. This example is for a viewpoint density of $3{\times }3$ viewpoints in the pupil. $d_p$ and $d_s$ refer to the diameters of the pupil and the subapertures, and ${\Delta }p$ and ${\Delta }q$ are the intervals of the subapertures in the $p$– and $q$–directions.
Fig. 3.
Fig. 3. Calculation and lateral shift of an elemental PSF. The diagrams show only the $x$ and $p$ dimension for simplicity. The red circle in the lens denotes the $n$-th subaperture and $p_n$ is the position of its centre. A) The geometry of a beam that emanates from the on-axis point in the CDP. The elemental PSF ($\mathrm {ePSF}(x,y;n,z_\mathrm {acc})$) is the retinal image of that point under a nominal accommodation distance of $z_\mathrm {acc}$. $A_\mathrm {elem}$ and $A_\mathrm {sub}$ indicate the amplitude variation of the incident beam and the subsequent amplitude modulation function for calculating the elemental PSF, respectively. B) The shifted elemental PSF when the LF display renders a point at the depth $z_r$. The shift $\Delta {}x_n$ is geometrically determined by $z_\mathrm {CDP}$, $z_\mathrm {eye}$, $z_r$, and $p_n$.
Fig. 4.
Fig. 4. Zernike coefficient values for modes 1–13 in OSA/ANSI order. Circles and error bars indicate mean and $\pm {}2$ standard deviation of each coefficient value in the statistical model. Colored lines represent generated Zernike coefficient sets for each virtual eye with a 3-mm pupil.
Fig. 5.
Fig. 5. A) The NCSF model by Watson & Ahumada (2005, 2012). The model is a two-dimensional real function that describes the neural portion of the contrast sensitivity function (CSF) as a filter. B) A cross section of the NCSF model, MTF, and CSF at vertical frequency = 0. Only half of one row is shown due to symmetry. The MTF (blue line) is calculated for the diffraction-limited eye with a 3-mm pupil. The CSF (black line) is a product of the MTF and the NCSF.
Fig. 6.
Fig. 6. Example of through-focus analysis for one open pupil of an aberrated eye. The CDP depth was 2D. OTFs are visualised in a manner that the brightness and hue correspondingly depict the OTF’s modulus and angle.
Fig. 7.
Fig. 7. Through-focus VSOTF for natural view. The grey curves indicate metric values as functions of relative accommodation, namely additional defocus to the rendered image depth. The grey vertical lines show their peaks. The red vertical line is the average of the peak positions over the eyes. The blue curve and vertical line represent the metric values and its peak for the average eye. The image depth was that of the CDP (2 dioptres).
Fig. 8.
Fig. 8. Through-focus VSOTF for the viewpoint densities of 2$\times$2, 3$\times$3, and 4$\times$4 viewpoints in a 3-mm pupil. Data are plotted in the same manner as Fig. 7. The green vertical lines indicate the CDP depth. Note that some parts of the through-focus ranges are missing in some panels on the top because we assumed the eyes would not accommodate hyperopically, namely its absolute accommodation distance should not be ‘farther than infinity’. Note that the scales of y-axes are not unified for better visibility of through-focus variations in each panel.
Fig. 9.
Fig. 9. Through-focus Strehl ratio for viewpoint densities of $2{\times }2$, $3{\times }3$, and $4{\times }4$ viewpoints in a 3-mm pupil. Data are shown in the same manner as Fig. 8.
Fig. 10.
Fig. 10. Through-focus VSOTF and Strehl ratio for an image rendered at the CDP depth $-2.0$D and PSFs and simulated retinal images at the two accommodation states in the through-focus range where the eye accommodates at the rendered image and the CDP. The PSFs and retinal images are for one of the aberrated eyes.
Fig. 11.
Fig. 11. Radial MTFs of the retinal images at predicted accommodation by VSOTF for each image depth and viewpoint density. The black plots with grey shades are the mean and $\pm {}1$ standard deviation of the radial MTFs for the aberrated eyes. The blue plots are the radial MTFs for the average eye.
Fig. 12.
Fig. 12. (A) Radial MTF for open 3-mm pupil at the in-focus state predicted by VSOTF. The blue line represents the radial MTF for the average eye. The black line and grey shaded area show the mean and $\pm$1 standard deviation of the radial MTFs for the aberrated eyes. (B) Cut-off frequencies for reference gain of 0.05 for the three viewpoint densities.

Equations (16)

Equations on this page are rendered with MathJax. Learn more.

[ Δ p Δ q ] = | z r z r z m | [ Δ u Δ v ] ,
A e l e m ( p , q ; n ) = | F r z m [ P m ( u , v ) F r z 0 [ δ ( s s n , t t n ) ] ] | ,
e P S F ( x , y ; λ , n , z a c c ) | F [ P s u b ( p , q ; λ , n , z a c c ) ] | ( x λ z e y e , y λ z e y e ) | 2
P s u b ( p , q ; λ , n , z a c c ) = A s u b ( p , q ; λ , n ) exp [ i 2 π λ W ( p , q ; λ , z a c c ) ] ,
A s u b ( p , q ; λ , n ) = A p u p i l ( p , q ) A S C E ( p , q ; λ ) A e l e m ( p , q ; n ) ,
e P S F p o l y ( x , y ; n , z a c c ) = e P S F ( x , y ; λ , n , z a c c ) V ( λ ) d λ .
Δ s n = [ Δ x n , Δ y n ] = z e y e ( 1 z r 1 z C D P ) [ p n , q n ] ,
P S F p o l y ( x , y ; z a c c ) = n e P S F p o l y ( x Δ x n , y Δ y n ; n , z a c c ) .
W ( p , q ) = m , n c n m Z n m ( p , q ) ,
D a c c ( z a c c ) = 1 z a c c 1 z C D P ,
c a c c ( z a c c ) = d p 2 16 3 D a c c ( z a c c ) \ ,
c a c c S A = 0.0435 z a c c .
D ( λ ) = A ( 1 λ c 1 λ r e f c ) ,
V S O T F = N C S F ( f x , f y ) O T F p o l y ( f x , f y ) d f x d f y N C S F ( f x , f y ) O T F D L ( f x , f y ) d f x d f y ,
O T F p o l y ( f x , f y ; z a c c ) = F [ P S F p o l y ( x , y ; z a c c ) ] ,
O T F D L ( f x , f y ) = F [ P S F p o l y D L ( x , y ) ] ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.