Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Megapixel photon-counting color imaging using quanta image sensor

Open Access Open Access

Abstract

Quanta Image Sensor (QIS) is a single-photon detector designed for extremely low light imaging conditions. Majority of the existing QIS prototypes are monochrome based on single-photon avalanche diodes (SPAD). Passive color imaging has not been demonstrated with single-photon detectors due to the intrinsic difficulty of shrinking the pixel size and increasing the spatial resolution while maintaining acceptable intra-pixel cross-talk. In this paper, we present image reconstruction of the first color QIS with a resolution of 1024 × 1024 pixels, supporting both single-bit and multi-bit photon counting capability. Our color image reconstruction is enabled by a customized joint demosaicing-denoising algorithm, leveraging truncated Poisson statistics andvariance stabilizing transforms. Experimental results of the new sensor and algorithm demonstrate superior color imaging performance for very low-light conditions with a mean exposure of as low as a few photons per pixel in both real and simulated images.

© 2019 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

Quanta image sensor (QIS) is a concept proposed by E. Fossum in 2005 aiming to overcome the limited sensor performance with the continuously shrinking pixel size [1, 2]. In a CMOS image sensor [3], images are acquired by accumulating hundreds to thousands of photoelectrons in each pixel and converting the charges to voltages and then to digital signals. As pixel size shrinks, the amount of photons that fall on individual pixels drops. Under low-light (or high-speed imaging) conditions, photon shot noise dominates the acquisition process. As the photon becomes scarce, eventually the signal-to-noise ratio will drop to a point that no meaningful images can be generated. Small pixel also reduces the full-well capacity which limits the dynamic range. QIS overcomes the problem by spatial-temporally oversampling the scene. In a QIS, each “image pixel” is partitioned into a group of smaller pixels called jots, and each jot is sensitive enough for photon counting. Over the past decade, numerous studies have demonstrated the feasibility of QIS, with a few prototypes developed [4–8]. On the image reconstruction side, various methods have also been proposed [9–12].

Despite the advances in QIS hardware and image reconstruction algorithms, one important open problem is the color imaging capability of QIS. Color imaging in low light is important for many computer vision applications such as night-vision. It is also important in biological studies, where using brighter light may lead to degradation of the biological matter. While active color imaging have been demonstrated using SPAD in [13] passive color imaging using QIS is difficult for two reasons. First, the resolution of a typical single-photon detector array is not high, e.g., 320 × 240 pixels [14], or most recently 512 × 512 pixels [8]. If a color filter array such as the standard Bayer pattern is implemented, the effective resolution will be halved to 160 × 120 or 256 × 256. For detection/tracking applications [15] this might be sufficient; but for photography the resolution could be low compared to the mainstream CMOS image sensors. Second, the color reconstruction of QIS data is different from CMOS image sensors because the QIS data is acquired under sparse-photon conditions with much lower signal-to-noise ratio. The demosaicing algorithm of QIS data needs toovercome the photon shot noise and accurately reproduce the color information. As we will discuss, QIS requires a joint reconstruction and demosaicing on top of the truncated Poisson statistics, a problem that has not been studied.

In this paper, we report the first demonstration of QIS color imaging. The QIS Pathfinder camera module developed at Gigajot Technology (as shown in Fig. 4) was used in this demonstration. The camera sensor has the same core design as the sensor described in [5], but has a Bayer pattern color filter array implemented on-chip. It has a spatial resolution of 1024 × 1024, single-bit and multi-bit photon-counting outputs, and up to 1040 fps speed in single-bit mode. Figure 1 shows a snapshot of the work presented in this paper. The key innovation, besides the sensor, is a new customized image reconstruction algorithm which enables joint reconstruction and demosaicing of raw QIS data. Our method leverages a few recently developed techniques in the QIS literature, including the transform-denoise method [10] and the Plug-and-Play Alternating Direction Method of Multipliers (ADMM) [16]. We demonstrate both our new sensor and the new algorithm on real data, and make quantitative evaluation based on synthetic data.

 figure: Fig. 1

Fig. 1 (a) One 1-bit frame. (b) Reconstructed color image using 50 frames of 1-bit input with threshold q = 4. (c) One 5-bit frame. (d) Reconstructed color image using 10 frames of 5-bit input. The average number ofphotons per frame is 5.

Download Full Size | PDF

2. Megapixel quanta image sensor

When the concept of QIS was first proposed, the two mainstream design methodologies to realize single-photon detection were the electron-multiplying charge-coupled device (EMCCD) [6] and the single-photon avalanche diode (SPAD) [17, 18]. Both devices amplify the signal from a single photoelectron with a physical phenomenon called electron avalanche multiplication, where high electrical voltage (typically higher than 20V) is used to accelerate the photoelectron to produce more free electrons through collision. SPAD is more widely adopted in consumer applications because of its excellent low-noise performance and temporal resolution. The Swiss SPAD [8, 19, 20] and the SPAD developed at University of Edinburgh [7, 14] are two better known examples. However due to its fabrication technology SPAD often has to make compromise in at least one of the three factors -low dark current, high quantum efficiency, and small pixel pitch. The QIS we use in this paper is based on a new type of single-photon detector proposed by Ma et al. [2, 4, 5, 21, 22]. The approach is to enhance the voltage signal generated by a single photoelectron by reducing the capacitance of the pixel output node so that the single-photon signal can overcome the background thermal noise.

 figure: Fig. 2

Fig. 2 Quanta Image Sensor (QIS) vs. CMOS Image Sensor (CIS). The top row shows a simulated CIS data at a photon level same as the QIS. The second row shows the real analog sensor data obtained from our prototype camera. For each photon flux level, we show both the raw input data and the denoised data. The CIS is assumed to have a read noise of 1.2e r.m.s., and we do not considered dark count. With dark count, the performance of CIS will deteriorate even more.

Download Full Size | PDF

Table 1 shows a comparison between a conventional CMOS image sensor, SPAD, and the prototype QIS used in this work. Among the listed different sensor parameters, the most noticeable difference between the QIS and SPADs is the small pixel pitch, where the QIS realized 1.1 μm pixel size, but SPADs typically have a pitch size around 5-20 μm. The smaller pixel size makes high spatial resolution (e.g. 10s or 100s of megapixel) possible with a small optical format (e.g. 1/4”). Another advantage of QIS over SPADs is the high quantum efficiency. Quantum efficiency measures the percentage of photons that are detectable by the sensor, which is critical for low-light imaging. SPAD cameras generally have a low quantum efficiency (e.g. <50%), because of low fill-factor and low photon detection capability normally called photon detection efficiency aka PDE. While low fill factor can be alleviated to an extent by using optical solutions such as microlens light-pipe. This reduces the photosensitive areal percentage in the SPAD pixels and limits the quantum efficiency. In contrast, since electron avalanche gain is not required in the QIS device, it has a quantum efficiency over 80% within the visible wavelengths.

Figure 2 shows comparison between real QIS image and a simulted CIS image. We note that we assume that the CIS no dark current. Even without considering the advantages of a lower dark current, the performance of QIS in photon starved regime is much better than the CIS.

Tables Icon

Table 1. Comparison of the available image sensor technologies.a

3. QIS color imaging model

The imaging model of a color QIS is illustrated in Fig. 3. Relative to the previous QIS imaging models, e.g., [10–12], the new model considers a color filter array which allows us to select wavelengths and multiple color channels. There are a few essential components in the model.

 figure: Fig. 3

Fig. 3 QIS Imaging Model. When the scene image arrives at the sensor, the color filter array first selects the wavelength according to the colors. Each color pixel is then sensed using a photon-detector. In single-bit mode the detector reports a binary value, and in multi-bit mode the detector reports an integer up to the saturation limit. The measured data contains three subsampled sequences, each representing a measurement in the color channel.

Download Full Size | PDF

Spatial oversampling. We represent the discretized scene image as a column vector c=[cr;cg;cb][0,1]3N, where N is the number of pixels, cr, cg, and cb are the light intensities of the red, green and blue channels, respectively. QIS was originally designed as a spatial-temporally oversampling device which samples each pixel using K jots spatially. This means every pixel has K measurements, and totally there are M=defNK measurements. Thus, the model of this oversampling process is

v=[vr;vg;vb]=αGc3M,
where G3M×3N represents the K-fold upsampling and interpolation filtering. The constant α is a gain factor. In our current prototype, the oversampling factor is K = 1 as we want to maximize the spatial resolution. In this case, we have G=I, and M=N.

Color filter array. To obtain color information, a color filter is placed on the top of each jot to collect light for only one of the RGB colors, and filter out the remaining two colors. This process is represented by three mutually exclusive M×M diagonal matrices or masks: Sr, Sg and SB for the red, green and blue channels, respectively. These matrices contain either zeros or ones on the diagonal, and Sr+Sg+Sb=I.

Incorporating the color filter array, the light exposure on the M jots of QIS is represented by

θ=Srvr+Sgvg+SbvbM.

We call θ the mosaiced image. If we let S=def[Sr,Sg,Sb]M×3M, then θ and c are related by

θ=αSGc.

 figure: Fig. 4

Fig. 4 Gigajot prototype QIS camera module and the QIS sensor chip. Note that there is no additional optics required besides a simple focusing lens. The camera is powered by standard 5V DC input, and has a USB3 data interface to transmit data to external storage.

Download Full Size | PDF

Photon arrival. The photon arrival of QIS is modeled as a Poisson process. Let YM be a non-negative random integer denoting the number of photons arrived at each jot during one integration period. Then, the probability of observing Y=yM is

(Y=y)=m=1Mθmymeθmym!.

For single-bit QIS, we put a voltage comparator to threshold the arrived photon count into a binary decision. Thus, the read out of each jot is a binary random variable B{0,1}M with Bm = 1 if Ymq and Bm = 0 if Ym<q, where q>0 is a threshold.

For multi-bit QIS, we use an L-bit quantizer to quantize the measurement. In this case,

Bm={Ym,Ym<2L1,2L1,Ym2L1.

In other words, we count the photons until the number reaches a threshold. When the threshold is reached, we report the maximum number before saturation.

 figure: Fig. 5

Fig. 5 Measured photo-electron count from a real sensor. The non-integer values of the count are due to the read noise of the readout circuit. However, with the ultra-low read noise provided by the QIS device, its negative effect on the photon-countingis negligible by setting appropriate thresholds.

Download Full Size | PDF

Assumptions. While the above model is theoretically tractable, we made three assumptions. The first is regarding the read noise associated with Y. If read noise is present, the measured photon count will become y+η where η is typically i.i.d. Gaussian. We assume that the read noise is negligible compared to the photon count, and can be eliminated by the thresholding process. We find that this is often the case with the QIS device used in this paper, which provides ultra-low read noise (0.24e- rms at room temperature) and accurate photon-counting capability. See Fig. 5 for an actual measurement of the photon count.

The second assumption we make is that the multi-bit saturation does not happen. That is, during the multi-bit sensing, we will never reach the upper threshold q and so Bm=Ym, which is a standard Poisson random variable. In terms of hardware, this can be achieved by implementing an automatic exposure control scheme so that the threshold q is increased when the scene is too bright. We will leave this to a follow up paper as it is outside the scope of this work.

The third assumption we make is that the camera is fast enough to catch up the scene movement, and the scene remains static during the multiple temporal samplings. This allows us to utilize multiple independent measurements over time to improve the statistics. In this case, the probability distribution of Y becomes

(Y=y)=t=1Tm=1Mθmym,teθmym,t!,
where ym,t is the Poisson count at pixel m in time t.

 figure: Fig. 6

Fig. 6 The image reconstruction pipeline consists of (i) a temporal binning step to sum the input frames, (ii) a variance stabilizing transform T to transform the measurement so that the variance is stabilized, (iii) a joint reconstruction and demosaicing algorithm to recovery the color, (iv) an inverse transform to compensate the forward transform, and (v) a tone mapping operation to correct the contrast.

Download Full Size | PDF

4. QIS color image reconstruction

The task of image reconstruction is to recover color scene c from the measurements B. In the gray-scale setting, we can formulate the problem as maximum-likelihood and solve it using convex optimization tools [9, 11, 12, 25]. We can also use learning-based methods, e.g., [26–29] to reconstruct the signal. The method we present here is based on the transform-denoise approach by Chan et al. [10]. We prefer transform-denoise because it is a physics-based approach and is robust to different sensor configurations. For example, in learning-based approaches, if we change the number of frames to sum, we need to train a different model or neural network.

4.1. Reconstruction pipeline

The pipeline of the proposed reconstruction algorithm is shown in Fig. 6. Given the random variable B={Bm,t} with realizations b={bm,t}, we first sum the frames to generate a single image Z={Zm}:

Zm=t=1TBm,t.

In the future, this step can be integrated into the hardware of the camera, so that the output of the camera will be the sum of multiple frames. Depending on whether we are using the single-bit mode or the multi-bit mode, the statistics of {Zm} follows either a Binomial distribution or a Poisson distribution. In single-bit mode, the distribution of {Zm} is

[Zm=k|θm]=(Tk)Ψq(θm)1k(1Ψq(θm))k,
where Ψq(θ)=def[B=0]=j=0q1θjeθ/j! is the probability of obtaining a zero in the binary variable B. As shown in [11],Ψq is the incomplete Gamma function. In multi-bit mode, since a sum of Poisson remains a Poisson, the distribution of {Zm} is
[Zm=k|θm]=(Tθm)keTθmk!.

If we assume that there is no processing of the bits besides temporal summation, then a simple maximum-likelihood of the probability will return us the solution:

θ=argmaxθ m=1M[Zm=zm|θm]=defM(z)={Ψq1(1z/T),Singlebit,z/T,Multibit.

The operation M can be interpreted as a tone-mapping.

4.2. Variance stabilizing transform

Now, we want to process the data besides just summing the frames. However, in either single-bit or multi-bit mode, the Binomial or the Poisson statistics are not easy as the variance changes with the mean. This prohibits the use any off-the-shelf algorithms that are based on i.i.d. Gaussian assumptions, e.g., denoising or inpainting.

The variance stabilizing transform (VST) is a statistical technique that tries to alleviate the changing variance problem [30–32]. Among the different types of VSTs, we are particularly interested in the two transforms originally proposed by F. Anscombe in 1948 [33]:

β=T(z)=def{T+12sin1z+38T+34,Singlebit,z+38,Multibit.

Here, the first transform is called the Anscombe Binomial, and the second is called the Anscombe Poisson. Both transforms are pixel-wise, meaning that we can apply the transform to each pixel independently. As proved in [10] and [33], the variance of T(z) is 1/4 for both the single-bit and the multi-bit case. The inverse T1 is simply the algebraic inverse (or the unbiased inverse if we follow [34]).

The important question for color imaging is now that the photon count z is the result of applying a color filter array matrix S=[Sr,Sg,Sb]. That is, instead of observing the full color z˜=def[zr;zg;zb], we only observe a subsampled version z=Sz˜. However, since T is point-wise and S is a concatenation of three non-overlapping binary matrices, it holds that

T(z)=T(Sz˜)=ST(z˜).

Consequently, if we define T(z) as the observed measurement (which is available from the sensor) and treat x=defT(z˜) as the unknown quantity to be determined, then the problem can be formulated as

v^=argminx SxT(z)2+λg(x),
where g is some regularization function controlling the smoothness of x. Note that (12) is a standard demosaicing-denoising problem assuming i.i.d. Gaussian noise.

4.3. Joint reconstruction and demosaicing

The optimization problem in (12) is a standard least squares with regularization function g. Thus, most convex optimization algorithm can be used as long as g is convex. In this paper, we adopt a variation of the alternating direction method of multiplier (ADMM) by replacing g with an off-the-shelf image denoiser. Such algorithm is coined the name Plug-and-Play (PnP) [35] (and different versions thereafter [16]).

For the particular problem in (12), the PnP ADMM algorithm iteratively updates the following two steps:

Demosaicing Module:

x(k+1)=(STS+ρI)1(STT(z)+ρ(v(k)u(k))),

Denoising Module:

v(k+1)=Dρ/λ(x(k+1)+u(k)),
and updates the Lagrange multiplier by u(k+1)=u(k)(x(k+1)v(k+1)). Readers interested in the detailed derivation can consult, e.g., [16]. Here, ρ is an internal parameter that controls the convergence. The operator D is an off-the-shelf image denoiser, e.g., Block-matching and 3D filtering (BM3D) or deep neural network denoisers. The subscript ρ/λ denotes the denoising strength, i.e., the hypothesized “noise variance”. Since STS is a diagonal matrix, the inversion is pointwise.

4.4. Non-iterative algorithm

The 1.1μm pixel pitch of the proposed QIS can potentially lead to a spatial resolution as high as or even higher than a conventional CMOS sensor. When this happens, in certain applications we can trade-off the color reconstruction efficiency and the resolution. For example, instead of using one jot for one pixel, we can use four jots for one pixel as shown in Fig. 7.

 figure: Fig. 7

Fig. 7 In the future when QIS can achieve higher spatial resolution, we can use four color jots to reconstruct one pixel. In this case, we can bypass the iterative ADMM algorithm and use a one-shot denoising method.

Download Full Size | PDF

Using four jots for one pixel allows us to bypass the iterative ADMM steps because there is no more missing pixel problem. In this case, the matrix SM×3M will become S=diag{14I,12I,14I}3M×3M, and hence (12) is simplified to a denoising problem with different noise levels for the three channels. In particular, the green channel has half of the variance of the red and the blue. For implementation, we can modify a denoiser, e.g., BM3D to accommodate the different noise variances. Since there is no more ADMM iteration, the algorithm is significantly faster. While we have used BM3D to demonstrate the results, any off the shelf denoiser which is used in CIS based cameras can be used to denoise the image for the four-jot to one-pixel method. We would also like to stress on the fact that both the Anscombe transform and the transform M can be implemented as a look up table. So, this method can be asfast a current denoiser being used in a CIS based camera.

4.5. Comparisons

We compare the proposed method with several existing methods on a synthetic dataset shown in Fig. 8. We simulate the raw color QIS data by assuming a Bayer pattern and using the image formation pipeline described in the previous section. We demosaic the images using: (a) a baseline method using MATLAB’s demosaic preceded by gray-scale BM3D denoising of R, G1, G2 and B channels and followed by color BM3D denoising; (b) Least-squares luma-chroma demultiplexing (LSLCD) method [36], which has a built-in BM3D denoiser; (c) Hirakawa’s PSDD method [37, 38], which does joint denoising and demosaicing for Poisson noise; and (d) the proposed method using BM3Dwith (λ,ρ)=(0.001,5). We apply variance stabilizing transform, except for PSDD which is designed for Poisson noise. The results show that the proposed method has a significantly better performance, both in terms of Peak Signal to Noise Ratio (PSNR) and visual quality.

 figure: Fig. 8

Fig. 8 Simulated QIS experiment. The goal of this experiment is to compare the proposed iterative algorithm with existing methods. We assume the observed Bayer RGB image is from a 3-bit QIS sensor. (a) Ground Truth; (b) One 3-bit QIS frame; (c) MATLAB demosaic preceded and followed by BM3D; (d) LSLCD [36]; (e) Hirakawa’s PSDD method [38], with a built-in wavelet shrinkage denoiser; (f) Proposed method with BM3D.

Download Full Size | PDF

5. Experimental results

5.1. Synthesized QIS data

We conduct a synthetic experiment to provide a quantitative evaluation of the performance of the proposed algorithm. To this end, we simulate the image formation pipeline by passing through color images to generate the QIS raw input data, with different number of bits. Figure 9 shows one example, and in the Supplementary Report we have additional examples.

In our simulation, we assume that the number of QIS frames is T = 4, and the average number of photon per pixel is 0.28, 0.85, 1.98 and 4.23 photons / frame for 1-bit, 2-bit, 3-bit and 4-bit QIS, respectively. On the measurement side, we generate single-bit and multi-bit data by thresholding the raw sensor output. To reconstruct the image, we use the proposed method with PnP and BM3D. The parameters are set as ρ = 1 and λ=0.007,0.003,0.002 and 0.0007 for 1-bit, 2-bit, 3-bit and 4-bit QIS, respectively. With as low as 1-bit, the reconstructed image in Fig. 9 is already capturing most of the features. As the number of bits increases, the visual quality improves.

 figure: Fig. 9

Fig. 9 Synthetic experiment for quantitative evaluation. [Top row]: One frame of the QIS measurements using different number of bits. [Bottom row]: Reconstructed images using the proposed method with 20 frames of QIS data. The average photon counts per pixel are 0.25, 0.75, 1.75 and 3.75 for 1-bit, 2-bit, 3-bit and 4-bit QIS, respectively.

Download Full Size | PDF

5.2. Real QIS data

We first show the reconstruction of an image of the Digital SG ColorChecker chart. We generate two sets of measurements: (a) a set of 50 one-bit frames, quantized with a threshold q = 4, and (b) a set of 10 five-bit frames. We use PnP and BM3D with (λ,ρ)=(0.15,10) and (0.01,10) for the 1-bit and 5-bit data, respectively. After reconstruction, the results are multiplied by a 3 × 3 color correction matrix to mitigate any color cross-talk. This matrix is generated by linear least square regression of the 24 Macbeth color patches that lies inside the SG ColorChecker chart. The results in Fig. 1 suggest that while 1-bit mode has color discrepancy with the ground truth, the 5-bit mode is producing a reasonably-high color accuracy.

Next, we show the result of imaging real scenes. See Fig. 10. In this experiment, the exposure time for each frame is set to 50μs. The average number of photons per pixel is approximately 4.2 photons for the ”QIS” sign image, 3 photons for the Pathfinder image, 1.9 photons for the duck image, and 2.9 photons for the mushroom image. For all images, we collect the data using a 5-bit QIS.

We demonstrate two algorithms: (i) the proposed transform-denoise framework using PnP + BM3D, and (ii) assuming four-jots to one-pixel scenario by trading half of the spatial resolution. The typical runtime on an unoptimized MATLAB code is approximately 4minutes for PnP, and 10 seconds for the four-jot to one-pixel method. An interesting observation is that even in the lower-resolution case, the details are not significantly deteriorated unless we zoom-in. However, the speed up we get is substantial.

 figure: Fig. 10

Fig. 10 Real QIS image reconstruction. The exposure time for each frame is 50 μs. The average number of photons per frame is 4.2, 3.0, 1.9, and 2.9 for each image respectively. Both methods use 4 frames for reconstruction. The raw data has a resolution of 1024 × 1024 pixels. The ADMM method retains the resolution, whereas the non-iterative method reduces the resolution to 512 × 512. Reconstruction using both the methods are shown at the same size for easier visual comparison. Notice that the non-iterative algorithm is able to achieve a visual quality almost similar to the ADMM method.

Download Full Size | PDF

6. Conclusion and discussion

We demonstrated the first color imaging for a mega-pixel Quanta Image Sensor (QIS). We designed a joint reconstruction-demosaicing algorithm for the quantized Poisson (and Poisson) statistics of the sensor data. Our solution involves a variance stabilizing transform and an iterative (and a non-iterative) algorithm. By integrating the sensor and the new algorithm, the sensor is able to acquire color images at a photon level as low as a few photons per pixel, a level that would be difficult for standard CMOS image sensors.

Funding

U.S. National Science Foundation Grant CCF-1718007.

Acknowledgement

The authors thank Prof. Eric Fossum for many valuable feedback. The 1Mjot color Quanta Image Sensor used in the demonstration was designed by J. Ma, S. Masoodian, and D. Starkey at Dartmouth College Advanced Image Sensor Laboratory led by Dr. Eric Fossum. The sensor was manufactured by Taiwan Semiconductor Manufacturing Company (TSMC). The camera module developed at Gigajot Technology was partially supported by National Science Foundation SBIR program. A Gnanasambandam, O. Elgendy and S. H. Chan are supported, in part, by the U.S. National Science Foundation under Grant CCF-1718007.

References

1. E. R. Fossum, “Some Thoughts on Future Digital Still Cameras,” in Image Sensors and Signal Processing for Digital Still Cameras, (CRC, 2006), 305–314.

2. E. R. Fossum, “Gigapixel digital film sensor (DFS) proposal,” in Nanospace Manipulation of Photons and Electrons for Nanovision Systems, The 7th Takayanagi Kenjiro Memorial Symposium and the 2nd International Symposium on Nanovision Science, Hamamatsu, Japan, (2005).

3. E. R. Fossum, “Active pixel sensors: Are CCDs dinosaurs?” in Charge-Coupled Devices and Solid State Optical Sensors III, vol. 1900 (International Society for Optics and Photonics, 1993), 2–15.

4. J. Ma and E. R. Fossum, “A pump-gate jot device with high conversion gain for a quanta image sensor,” IEEE J. Electron Devices Soc. 3, 73–77 (2015).

5. J. Ma, S. Masoodian, D. A. Starkey, and E. R. Fossum, “Photon-number-resolving megapixel image sensor at room temperature without avalanche gain,” Optica 4, 1474–1481 (2017). [CrossRef]  

6. J. Hynecek, “Impactron-a new solid state image intensifier,” IEEE Trans. Electron Devices 48, 2238–2241 (2001). [CrossRef]  

7. N. A. W. Dutton, L. Parmesan, A. J. Holmes, L. A. Grant, and R. K. Henderson, “320×240 oversampled digital single photon counting image sensor,” in Proceedings of IEEE 2014 Symposium on VLSI Circuits Digest of Technical Papers, (IEEE, 2014).

8. C. Bruschini, S. Burri, S. Lindner, A. C. Ulku, C. Zhang, I. M. Antolovic, M. Wolf, and E. Charbon, “Monolithic SPAD arrays for high-performance, time-resolved single-photon imaging,” in Proceedings of International Conference on Optical MEMS and Nanophotonics, (IEEE, 2018).

9. S. H. Chan and Y. M. Lu, “Efficient image reconstruction for gigapixel quantum image sensors,” in Proceedings of IEEE Global Conference on Signal and Information Processing, (IEEE, 2014), 312–316.

10. S. H. Chan, O. A. Elgendy, and X. Wang, “Images from bits: Non-iterative image reconstruction for quanta image sensors,” Sensors 16, 1961 (2016). [CrossRef]   [PubMed]  

11. O. A. Elgendy and S. H. Chan, “Optimal threshold design for quanta image sensor,” IEEE Trans. Comput. Imag. 4, 99–111 (2018).

12. F. Yang, Y. M. Lu, L. Sbaiz, and M. Vetterli, “Bits from photons: Oversampled image acquisition using binary Poisson statistics,” IEEE Trans. Image Process. 21, 1421–1436 (2012).

13. X. Ren, Y. Altmann, R. Tobin, A. Mccarthy, S. Mclaughlin, and G. S. Buller, “Wavelength-time coding for multispectral 3d imaging using single-photon lidar,” Opt.Express 26, 30146–30161 (2018).

14. N. A. Dutton, I. Gyongy, L. Parmesan, and R. K. Henderson, “Single photon counting performance and noise analysis of CMOS SPAD-based image sensors,” Sensors 161122 (2016). [CrossRef]   [PubMed]  

15. I. Gyongy, N. A. Dutton, and R. K. Henderson, “Single-photon tracking for high-speed vision,” Sensors 18323 (2018). [CrossRef]  

16. S. H. Chan, X. Wang, and O. A. Elgendy, “Plug-and-play ADMM for image restoration: Fixed-point convergence and applications,” IEEE Trans. Comput. Imag. 384–98 (2017).

17. B. F. Aull, A. H. Loomis, D. J. Young, R. M. Heinrichs, B. J. Felton, P. J. Daniels, and D. J. Landers, “Geiger-mode avalanche photodiodes for three-dimensional imaging,” Linc. Lab. J. 13, 335–349 (2002).

18. J. Buchholz, J. Krieger, C. Bruschini, S. Burri, A. Ardelean, E. Charbon, and J. Langowski, “Widefield high frame rate single-photon SPAD imagers for SPIM-FCS,” Biophys. J. 114, 2455–2464 (2018). [CrossRef]   [PubMed]  

19. E. Charbon, “Towards large scale CMOS single-photon detector arrays for lab-on-chip applications,” J. Phys. D: Appl. Phys. 41, 094010 (2008). [CrossRef]  

20. E. Charbon, “Will avalanche photodiode arrays ever reach 1 megapixel,” in Proceedings of International Image Sensor Workshop, (IISS, 2007), 246–249.

21. J. Ma and E. R. Fossum, “Quanta image sensor jot with sub 0.3 e-rms read noise and photon counting capability,” IEEE Electron Device Lett. 36, 926–928 (2015). [CrossRef]  

22. S. Masoodian, J. Ma, D. Starkey, Y. Yamashita, and E. R. Fossum, “A 1Mjot 1040fps 0.22 e-rms stacked BSI quanta image sensor with Cluster-Parallel Readout,” in Proceedings of the International Image Sensor Workshop, vol. 30 (IISS, 2017), 230–233.

23. N. A. Dutton, I. Gyongy, L. Parmesan, S. Gnecchi, N. Calder, B. R. Rae, S. Pellegrini, L. A. Grant, and R. K. Henderson, “A SPAD-based QVGA image sensor for single-photon counting and quanta imaging,” IEEE Trans. Electron Devices 63, 189–196 (2016). [CrossRef]  

24. “Zylus 4.2 PLUS,” https://andor.oxinst.com/products/scmos-camera-series/zyla-4-2-scmos. Accessed: 2019-04-21.

25. F. Yang, Y. M. Lu, L. Sbaiz, and M. Vetterli, “An optimal algorithm for reconstructing images from binary measurements,” in Computational Imaging VIII, vol. 7533 (International Society for Optics and Photonics, 2010), 75330K. [CrossRef]  

26. T. Remez, O. Litany, and A. Bronstein, “A picture is worth a billion bits: Real-time image reconstruction from dense binary threshold pixels,” in Proceedings of IEEE International Conference on Computational Photography, (IEEE, 2016), 1–9.

27. J. H. Choi, O. A. Elgendy, and S. H. Chan, “Image reconstruction for quanta image sensors using deep neural networks,” in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), (IEEE, 2018), 6543–6547.

28. P. Chandramouli, S. Burri, C. Bruschini, E. Charbon, and A. Kolb, “A little bit too much? High speed imaging from sparse photon counts,” arXiv preprint arXiv:1811.02396 (2018).

29. R. A. Rojas, W. Luo, V. Murray, and Y. M. Lu, “Learning optimal parameters for binary sensing image reconstruction algorithms,” in Proceedings of IEEE International Conference on Image Processing, (IEEE, 2017), 2791–2795.

30. L. Azzari and A. Foi, “Variance stabilization in Poisson image deblurring,” in IEEE International Symposium on Biomedical Imaging, (IEEE, 2017), 728–731.

31. L. Azzari and A. Foi, “Variance stabilization for noisy+ estimate combination in iterative Poisson denoising,” IEEE Signal Process. Lett. 23, 1086–1090 (2016). [CrossRef]  

32. A. Foi, “Noise estimation and removal in MR imaging: The variance-stabilization approach,” in Proceedings of International symposium on biomedical imaging: from nano to macro, (IEEE, 2011), 1809–1814.

33. F. J. Anscombe, “The transformation of Poisson, binomial and negative-binomial data,” Biometrika 35, 246–254 (1948). [CrossRef]  

34. M. Makitalo and A. Foi, “Optimal inversion of the Anscombe transformation in low-count Poisson image denoising,” IEEE Trans. Image Process. , 20, 99–109 (2011). [CrossRef]  

35. S. H. Chan, “Performance analysis of plug-and-play admm: A graph signal processing perspective,” IEEE Trans. Comput. Imag. (2019). [CrossRef]  

36. G. Jeon and E. Dubois, “Demosaicking of noisy Bayer-sampled color images with least-squares Luma-Chroma demultiplexing and noise level estimation,” IEEE Trans. Image Process. 22, 146–156 (2013).

37. J. T. Korneliussen and K. Hirakawa, “Camera processing with chromatic aberration,” IEEE Trans. on Image Process. 23, 4539– 4552 (2014). [CrossRef]  

38. K. Hirakawa, X. Meng, and P. J. Wolfe, “A framework for wavelet-based analysis and processing of color filter array images with applications to denoising and demosaicing,” in IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 1 (IEEE, 2007), 597–600.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (10)

Fig. 1
Fig. 1 (a) One 1-bit frame. (b) Reconstructed color image using 50 frames of 1-bit input with threshold q = 4. (c) One 5-bit frame. (d) Reconstructed color image using 10 frames of 5-bit input. The average number ofphotons per frame is 5.
Fig. 2
Fig. 2 Quanta Image Sensor (QIS) vs. CMOS Image Sensor (CIS). The top row shows a simulated CIS data at a photon level same as the QIS. The second row shows the real analog sensor data obtained from our prototype camera. For each photon flux level, we show both the raw input data and the denoised data. The CIS is assumed to have a read noise of 1.2 e r.m.s., and we do not considered dark count. With dark count, the performance of CIS will deteriorate even more.
Fig. 3
Fig. 3 QIS Imaging Model. When the scene image arrives at the sensor, the color filter array first selects the wavelength according to the colors. Each color pixel is then sensed using a photon-detector. In single-bit mode the detector reports a binary value, and in multi-bit mode the detector reports an integer up to the saturation limit. The measured data contains three subsampled sequences, each representing a measurement in the color channel.
Fig. 4
Fig. 4 Gigajot prototype QIS camera module and the QIS sensor chip. Note that there is no additional optics required besides a simple focusing lens. The camera is powered by standard 5V DC input, and has a USB3 data interface to transmit data to external storage.
Fig. 5
Fig. 5 Measured photo-electron count from a real sensor. The non-integer values of the count are due to the read noise of the readout circuit. However, with the ultra-low read noise provided by the QIS device, its negative effect on the photon-countingis negligible by setting appropriate thresholds.
Fig. 6
Fig. 6 The image reconstruction pipeline consists of (i) a temporal binning step to sum the input frames, (ii) a variance stabilizing transform T to transform the measurement so that the variance is stabilized, (iii) a joint reconstruction and demosaicing algorithm to recovery the color, (iv) an inverse transform to compensate the forward transform, and (v) a tone mapping operation to correct the contrast.
Fig. 7
Fig. 7 In the future when QIS can achieve higher spatial resolution, we can use four color jots to reconstruct one pixel. In this case, we can bypass the iterative ADMM algorithm and use a one-shot denoising method.
Fig. 8
Fig. 8 Simulated QIS experiment. The goal of this experiment is to compare the proposed iterative algorithm with existing methods. We assume the observed Bayer RGB image is from a 3-bit QIS sensor. (a) Ground Truth; (b) One 3-bit QIS frame; (c) MATLAB demosaic preceded and followed by BM3D; (d) LSLCD [36]; (e) Hirakawa’s PSDD method [38], with a built-in wavelet shrinkage denoiser; (f) Proposed method with BM3D.
Fig. 9
Fig. 9 Synthetic experiment for quantitative evaluation. [Top row]: One frame of the QIS measurements using different number of bits. [Bottom row]: Reconstructed images using the proposed method with 20 frames of QIS data. The average photon counts per pixel are 0.25, 0.75, 1.75 and 3.75 for 1-bit, 2-bit, 3-bit and 4-bit QIS, respectively.
Fig. 10
Fig. 10 Real QIS image reconstruction. The exposure time for each frame is 50 μs. The average number of photons per frame is 4.2, 3.0, 1.9, and 2.9 for each image respectively. Both methods use 4 frames for reconstruction. The raw data has a resolution of 1024 × 1024 pixels. The ADMM method retains the resolution, whereas the non-iterative method reduces the resolution to 512 × 512. Reconstruction using both the methods are shown at the same size for easier visual comparison. Notice that the non-iterative algorithm is able to achieve a visual quality almost similar to the ADMM method.

Tables (1)

Tables Icon

Table 1 Comparison of the available image sensor technologies.a

Equations (15)

Equations on this page are rendered with MathJax. Learn more.

v = [ v r ; v g ; v b ] = α Gc 3 M ,
θ = S r v r + S g v g + S b v b M .
θ = α SGc .
( Y = y ) = m = 1 M θ m y m e θ m y m ! .
B m = { Y m , Y m < 2 L 1 , 2 L 1 , Y m 2 L 1.
( Y = y ) = t = 1 T m = 1 M θ m y m , t e θ m y m , t ! ,
Z m = t = 1 T B m , t .
[ Z m = k | θ m ] = ( T k ) Ψ q ( θ m ) 1 k ( 1 Ψ q ( θ m ) ) k ,
[ Z m = k | θ m ] = ( T θ m ) k e T θ m k ! .
θ = argmax θ   m = 1 M [ Z m = z m | θ m ] = def M ( z ) = { Ψ q 1 ( 1 z / T ) , Single bit , z / T , Multi bit .
β = T ( z ) = def { T + 1 2 sin 1 z + 3 8 T + 3 4 , Single bit , z + 3 8 , Multi bit .
T ( z ) = T ( S z ˜ ) = S T ( z ˜ ) .
v ^ = argmin x   Sx T ( z ) 2 + λ g ( x ) ,
x ( k + 1 ) = ( S T S + ρ I ) 1 ( S T T ( z ) + ρ ( v ( k ) u ( k ) ) ) ,
v ( k + 1 ) = D ρ / λ ( x ( k + 1 ) + u ( k ) ) ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.