Abstract
We present a feature-specific imaging system based on the use of structured illumination. The measurements are defined as inner products between the illumination patterns and the object reflectance function, measured on a single photodetector. The illumination patterns are defined using random binary patterns and thus do not employ prior knowledge about the object. Object estimates are generated using L 1-norm minimization and gradient-projection sparse reconstruction algorithms. The experimental reconstructions show the feasibility of the proposed approach by using 42% fewer measurements than the object dimensionality.
©2008 Optical Society of America
1. Introduction
Imaging is employed for various applications within both defense and commercial domains. Depending on the application or task, imaging modalities either assume ambient illumination or rely upon active/structured illumination. Active illumination is valuable for various applications like laser radar, digital holography, 3D imaging (e.g., depth-map extraction) and microscopy (e.g., grating-based superresolution) [1–7]. The measurement space associated with such systems is in general, not an isomorphism of the object space and computation plays an important role in performing the associated task. Recently a great deal of interest has arisen in so-called computational imagers that are optimized towards a certain system goal [8–12]. Here we are interested in a form of computational imaging that is compressive or feature-specific in nature. Feature-specific imaging (FSI) measures linear projections of the object space optically and can exploit these feature measurements for reconstruction, compression and/or recognition tasks [12–16]. The projection basis is designed such that it can extract the task-related information from the scene using fewer measurements than a conventional imager. Significant benefits in terms of improved reconstruction fidelity, improved recognition performance and improved measurement signal-to-noise-ratio have been reported in the literature [12–16].
It is possible to extend the capabilities of FSI by exploiting illumination degrees of freedom. Such a feature-specific structured imaging (FSSI) system has been described in [17]. FSSI projects spatially structured illumination patterns onto an object and collects the reflected light onto a single photo-detector resulting in feature measurements. It was demonstrated that FSSI may offer benefits in terms of reduced number of measurements and hardware complexity compared to a conventional imager [17]. This previous work on FSSI makes use of detailed object prior knowledge to define the illumination patterns. In this work we generalize FSSI by making use of random projections to define the illumination patterns. Several reports exist on the use of random projections (RP) for sparse signal/image recovery [14, 18–20]. The main appeal behind using RP is that one does not require detailed prior knowledge about the object during the measurement process. Instead, sparsity (a form of object prior knowledge) is enforced during the reconstruction process through nonlinear algorithms based on L 1-norm minimization and/or gradient-projection [18–20]. The benefits of the present work over the passive compressive imaging architecture reported in [14] are twofold: (1) the active modality enables imaging with zero ambient light levels, and (2) the active modality reduces the complexity of the light-collection hardware. The advantage of this work compared to our previous FSSI report [17] is that the measurement mechanism uses a RP basis and so does not depend on the object. We note however that the reconstruction fidelity is dependent on the incoherence between the RP basis and the basis in which the object is sparse [18, 19].
The remainder of our paper is organized as follows. In section 2 we describe the RFSSI system. Section 3 presents the experimental setup along with results. Section 4 offers concluding remarks.
2. Imager system analysis
Figure 1 depicts the RFSSI system. It consists of 1) an illuminator to project a sequence of spatially structured intensity patterns onto a two-dimensional object, 2) light collection optics focussing the reflected light onto a single photodetector and 3) a post-processing algorithm to generate an object estimate from the sequence of measurements. We assume that 1) the illuminator resolution is matched to that of the object i.e. it provides resolution at least equal to the object pixel size, and 2) the illuminator and the light collection optics share a common optical axis. The object reflectance function is of dimension on N×N and is lexicographically ordered into a N 2-dimensional vector G. All two-dimensional quantities will be represented in this same manner throughout the paper. The illumination pattern is considered to be spatially incoherent and is represented by the vector P. Because the illuminator and light collection optics are assumed to be in a mono-static configuration the reflected irradiance pattern from the object will be a pointwise product between P and G. This reflected light is focused onto a single photodetector resulting in the feature measurement given by
The feature measurement R in Eq. (1) is simply an inner product of P and G. We now generalize the measurement model in Eq. (1) by considering a sequence of K projections. The resulting K -dimensional measurement vector R is given by
where P̿ represents the K×N 2 measurement matrix, P̿=[P 1,P 2……P K]T where P i is the ith projection vector of dimension N 2×1.
Before describing the postprocessing algorithm it is important to discuss the assumptions related to object sparsity. The object reflectance vector G can be represented as a N 2- dimensional real-valued vector α in an orthonormal basis V̿ (of dimension N 2×N 2) so that G=V̿α. We will refer to V̿ as the representation matrix and it can be expressed as where for i=j and equal to 0 otherwise. By sparsity of the object reflectance G we mean that the number of non-zero elements (M) in the vector α is much smaller than its dimensionality i.e. M≪N 2. We will refer to such an object as “explicitly-sparse.” Consider the example in Fig. 2(a) where we show two different 32×32 binary sparse objects with M=160 non-zero values. The representation basis V̿ in this case is simply an identity matrix so that α=G and therefore these objects are “explicitly-sparse” in the space domain. However, there are scenarios where the coefficient vector α is not explicitly sparse but still compressible. This implies that α has only M≪N 2 elements larger than ε in magnitude, where ε≪max(|α|). We will refer to such objects as “mostly-sparse.” Consider the example in Fig. 2(b) where we show two different 32×32 gray-scale truck objects (reflectance normalized to 1) that are “mostly-sparse” in a wavelet basis. The wavelet basis V̿ DW is created using the Daubechies’ wavelet filters of length equal to 4 with number of levels equal to 7 [21]. Figure 2(c) plots the coefficient vector α representing the leftmost object from Fig. 2(b) in the basis V̿ DW. Note from Fig. 2(c) that max(|α|)=7.09 and the number of elements greater than ε=0.18 is equal to 250. We create a truncated coefficient vector α trunc by retaining these dominant 250 coefficients of α at their respective positions and equating the remaining elements to zero. The truncated coefficient vector α trunc is transformed back to the object space to form G trunc=V̿ DW α trunc which is shown in Fig. 2(d). Note that the visual quality remains good when retaining only 250 coefficients of the wavelet expansion. We can observe that the object in Fig. 2(d) is now explicitly sparse in the wavelet domain.
Next, we consider the design of the measurement matrix and the associated postprocessing required for the RFSSI system discussed above. The postprocessing algorithm is to be designed such that it yields an estimate (Ĝ) of the object reflectance (G). The reconstruction quality is evaluated using the root-mean-squared-error (RMSE) criterion:
where denotes the L 2-norm of the real-valued vector x and is given by . Here we are interested in the case where the number of measurements K≪N 2 and therefore solving for Ĝ becomes an underdetermined problem. We use the result by Candes et al. in [18] where it is shown that if (1) G is M-sparse, (2) the elements of P̿ are independently chosen from the symmetric Bernoulli distribution Pr(P̿ ij=0 or 1)=1/2 and, (3) K M·log(N 2/M) then G can be reconstructed exactly with high probability by solving the convex program
Note that denotes the L 1-norm of the real-valued vector x and is given by . We will refer to this approach as L1. Candes’ result essentially says that the measurement process need not rely on detailed object prior knowledge. The “cost” of our ignorance will be the need to take roughly log(N 2/M) times more measurements as compared to the case in which object prior knowledge (i.e., the specific directions containing maximum object information) is available during the measurement. Next, we will present simulation results to demonstrate the operation of the reconstruction algorithm in Eq. (4). Figure 3(a) shows a 32×32 binary object reflectance pattern with M=160 non-zero elements. Therefore, from the above mentioned result of Candes, we would need K≥log(N 2/M)=300 measurements to obtain RMSE=0. We use L1 to reconstruct the reflectance estimate and fig. 3(b) shows the estimate Ĝ for K=350 and K=450. The solid curve with ‘circles’ in Fig. 3(e) shows the resulting reconstruction RMSE, for the binary object in Fig. 3(a), as a function of K. The RMSE trend with respect to K is monotonically decreasing eventually going to zero. Note that the resulting RMSE values for K=350 and K=450 are 0.133 and 0 respectively. Figure 3(c) shows a 32×32 gray-scale object that is “mostly-sparse” in the wavelet basis V̿ DW. Although this object is not explicitly sparse in the V̿ DW basis we can still apply the L1 algorithm because of its “mostly-sparse” nature. Figure 3(d) shows the estimate Ĝ for K=350 and K=450. Note that the resulting RMSE values for K=350 and K=450 are 0.19 and 0.164 respectively. The solid curve with ‘squares’ in Fig. 3(e) represents the reconstruction RMSE for this “mostly-sparse” object.
Note that the reconstruction is not exact because the object is not explicitly-sparse as required by the Candes’ result in Eq. (4).
We will now extend the model in Eq. (2) to the case where our observations are noisy. Consider the new measurement model for RFSSI given by
where n represents the K-dimensional error/perturbation vector. We will assume that this perturbation term has bounded energy i.e. . For example, if each element of n is a zero-mean white Gaussian (WG) random variable with variance σ 2 o then σ=σ o √K. Candes et al. extends the L1 approach in [19] by incorporating knowledge about the noise. The estimate Ĝ can now be obtained by solving the modified convex program
Candes et al. solve this convex program by recasting it as a second order cone program (SOCP) [19]. Recently Figueiredo et al. has proposed a gradient-projection based approach to solve a slight variant of the problem in Eq. (6) [20]. This approach attempts to minimize the weighted sum of the data fidelity term and the L 1-norm under no constraints. Figueiredo’s approach is shown to be computationally faster than SOCP in [20] with no sacrifice in the reconstruction quality. This method is referred to as gradient projection for sparse reconstruction (GPSR) in [20]. We make use of the MATLAB routine for GPSR publicly available at www.lx.it.pt/~mtf/GPSR. We now consider a simulation example based on the use of GPSR. The error term n in Eq. (5) will be modeled as additive WG noise with σo=2 which is equivalent to 7% of the dynamic range of the corresponding error-free measurement in Eq. (2). The measurement matrix P̿ is created using random binary patterns as explained above. Figure 4(a) shows the estimates Ĝ of the object in fig. 3(a) for K=450 and K=600. The resulting RMSE for this object is plotted in Fig. 4(b). Observe that RMSE is monotonically decreasing with respect to K. As opposed to the trend in Fig. 3(c) RMSE does not converge to zero due to measurement noise. Using the knowledge that M=160 we retain the dominant 160 elements of the estimates Ĝ in Fig. 4(a) resulting in the modified estimates shown in Fig. 4(c). This approach is however not optimal and therefore, we did not compare it with the GPSR method.
3. Experimental setup
Figure 5 depicts the experimental setup of the RFSSI system. The illuminator used for the experiment is an Epson PowerLite S1+ projector. It is controlled by a computer that provides a sequence of random binary-valued images. The light collection optics is implemented using a f/1 biconvex lens with aperture size equal to 25.4 mm. The photodetector used in the experiment is a Newport 818-SL connected to a Newport 1830-C power meter. The integration time used for each feature measurement is equal to 250 ms. Both the illuminator and light collection optics are placed at a distance of 95 cm from the 2D object. The 2D object has 32×32 pixels with scene dimensions equal to 18 cm by 18 cm, and is implemented through a poster-board image. Therefore, the scene defines a 10.7° field of view as seen from the detector. The physical pixel size of the object is 5.6 mm and is larger than the projected pixel size which is roughly 3 mm.
Before describing the calibration procedure we discuss the kinds of sparse objects (of dimension 32×32; N 2=1024) being used for testing purposes. For experimental validation of the proposed RFSSI system we consider three kinds of objects: 1) Binary objects (“explicitly-sparse” in the space domain) with M=160, 2) Gray-scale truck objects “explicitly-sparse” in a principal component (PC) basis with M=160, and 3) Gray-scale truck objects “mostly-sparse” in a PC or wavelet basis. We have already explained the first and the third kind of objects in Section 2. Here we describe the procedure of generating objects that are “explicitly-sparse” in a PC basis. The PC basis is determined by the statistics of an object ensemble [21]. For a set of objects S, the PC projections are defined as the eigenvectors of the object autocorrelation matrix RS=E{GG T}, where G belongs to S and E{·} denotes the expectation operator. Here we consider 4000 truck objects to define S and subsequently generate the PC basis V̿ PC. Each object reflectance G will therefore be represented as a N 2- dimensional real-valued vector α in the PC basis V̿ PC (of dimension N 2×N 2) i.e. G=V̿ PC α. The “explicitly-sparse” objects (with M=160) in the PC basis can thus be generated by retaining only the largest 160 coefficients of |α|. Figure 6(a) shows two example truck objects used for generating Rs and fig. 6(b) shows their respective “explicitly-sparse” (with M=160) versions.
Next, we describe the calibration procedure used in our experiment. The idea is to perform curve-fitting between the theoretical and experimental feature measurements for a set of five known training objects. Figure 7(a) shows the five training objects. The number of measurements K for each training object is taken to be 600 i.e. the measurement matrix P̿ is of dimension 600×1024. The K-dimensional theoretical measurement vector is given by R in Eq. (2). We will denote the K-dimensional experimental measurement vector as R exp. Figure 7 (b) plots the first 100 entries of R and R exp corresponding to the leftmost object in Fig. 7(a).
We fit an affine-linear model given by R calib=a·R exp+b such that the root-least-squares error for each training object is minimized. The values for a and b determined using the training data are 0.93 and -157.7 respectively. The resulting residual error after fitting the model is found to be RLSE≈2. This residual error is equivalent to 7% of the dynamic range of the theoretical measurement vector R, and can be modeled as the perturbation term defined in Eq. (5) because it is bounded. Various factors that could be contributing to this error are detector noise or read-out noise, illuminator temperature variations and improper reflection model. Figure 7(c) plots the first 100 entries of R and R calib corresponding to the leftmost object in Fig. 7(a).
Now we will use the calibration parameters obtained above to generate experimental reconstructions for several test objects. The steps involved in experimental reconstruction are: 1) measure the readings associated with K (equal to 600 for this study) illumination patterns defined by P̿ and store them in vector R exp, 2) transform R exp into R calib by using the calibration parameters derived above and, 3) use the GPSR approach to generate the object estimate Ĝ from R calib. Figure 8(a) shows two binary test objects along with their respective estimates obtained from the experiment. Figure 8(b) shows two explicitly-sparse objects in the V̿ PC basis along with their respective estimates. Figure 8(c) shows two “mostly-sparse” truck objects in the V̿ PC basis along with their respective estimates. Figure 8(d) shows two “mostly-sparse” truck objects in the V̿ DW basis along with their respective estimates. These experimental reconstructions demonstrate the feasibility of RFSSI by using only 600 measurements i.e. 58% of the object dimensionality. The experimental RMSE, obtained by averaging over the test objects in fig. 8, for K=600 (or equivalently 58%) is equal to 0.17. Figure 9 shows the RMSE performance, averaged over the test objects in fig. 8, as a function of the number of feature measurements K. The curve with circles represents theory and the curve with squares represents our experiment performance. The data in fig. 9 shows excellent agreement between experiment and simulation.
Lastly we will compare the RFSSI and FSSI [17] systems in terms of the number of measurements required to achieve a specified RMSE. The number of measurements required by RFSSI and FSSI to achieve a specified RMSE will be denoted by KRFSSI and KFSSI respectively. As noted earlier in Section 1, FSSI makes use of substantial object prior knowledge (i.e. object autocorrelation matrix RS and PC basis V̿ PC) during the measurement process. From [17] we find that the reconstruction RMSE for FSSI is lower-bounded by
where Tr{·} denotes the trace operator, P̿ PC denotes the measurement matrix (of dimension KFSSI×N 2) and [I] denotes an identity matrix of dimension KFSSI×KFSSI. The measurement matrix P̿ PC is created by retaining the KFSSI dominant eigenvectors from V̿ PC. We will denote the experimental RMSE performance of the RFSSI system, averaged over the gray-scale test objects, as RMSERFSSI. To ensure a fair comparison between RFSSI and FSSI, we make sure that both systems use the same illumination energy for each pattern. This is done by enforcing for i =1…N 2. Table 1 compares KRFSSI and KFSSI for several values of RMSE. It can be seen that KFSSI≪KRFSSI as expected.
4. Conclusions
We have presented a feature-specific structured imaging system that uses binary random projections to define the illumination patterns. The feature measurements represent the inner products of the illumination patterns with the object reflectance and are measured optically onto a single photodetector. The object reflectance estimates were generated by using nonlinear algorithms based on L 1-norm minimization and/or gradient-projection sparse reconstruction. We consider two kinds of “explicitly-sparse” and “mostly-sparse” objects for testing the RFSSI system. We have presented an experimental validation of the RFSSI approach. The experimental reconstructions utilize only 600 measurements i.e. 58% of the object dimensionality.
Acknowledgements
We gratefully acknowledge the financial support of DARPA’S MONTAGE Program.
References
1. Q. Zheng, S. Der, and H. I. Mahmoud, “Model-based target recognition in pulsed ladar imagery,” IEEE Trans. Image Process. 10, 565–572 (2001). [CrossRef]
2. P. Wheel, M. Dobbs, and W. E. Sharp, “Optimization of space borne imaging LADAR sensor for asteroid studies from parameter design,” in Electro-Optical System Design, Simulation, Testing, and Training, R. M. Wasserman and S. L. DeVore, eds., Proc. SPIE4772, 68–77 (2002). [CrossRef]
3. S. Lai, B. King, and M. A. Neifeld, “Wave front reconstruction by means of phase-shifting digital in-line holography,” Opt. Commun. 173, 155–160 (2000). [CrossRef]
4. J. Batlle, E. Mouaddib, and J. Salvi, “Recent progress in coded structured light as a technique to solve the correspondence problem: a survey,” Pattern Recogn. 31, 963–982 (1998). [CrossRef]
5. E. Horn and N. Kiryati, “Toward optimal structured light patterns,” Image Vision Comput. 17, 87–97 (1999). [CrossRef]
6. M.G.L. Gustafsson, “Surpassing the lateral resolution limit by a factor of two using structured illumination microscopy,” J. Microsc. 198, 82–87 (2000). [CrossRef] [PubMed]
7. J. Ryu, B.K.P. Horn, M.S. Mermelstein, S. Hong, and D.M. Freeman, “Application of Structured Illumination in Nano-Scale Vision,” in IEEE Workshop on Computer Vision for the Nano-Scale, Madison, Wisconsin, 16–22 (2003).
8. W. T. Cathey and E. R. Dowsky, “New paradigm for imaging systems,” Appl. Opt. 41, 6080–6092 (2002). [CrossRef] [PubMed]
9. S. Prasad, T. C. Torgersen, V. P. Pauca, R. J. Plemmons, and J. van der Gracht, “Engineering the pupil phase to improve image quality,” in Visual Information Processing XII, Z. Rahman, R. Schowengerdt, and S. Reichenbach, eds., Proc. SPIE5108, 1–12 (2003). [CrossRef]
10. P. Potuluri, M.R. Fetterman, and D.J. Brady, “High depth of field microscopic imaging using an interferometric camera,” Opt. Express 8, 624–630 (2001). [CrossRef] [PubMed]
11. D. J. Brady, “Multiplex sensors and the constant radiance theorem,” Opt. Lett. 27, 16–18 (2002). [CrossRef]
12. M. A. Neifeld and P. Shankar, “Feature-specific imaging,” Appl. Opt. 42, 3379–3389 (2003). [CrossRef] [PubMed]
13. N. P. Pitsianis, D. J. Brady, and X. Sun, “The Quantized Cosine Transform for sensor-layer image compression,” in Adaptive Optics: Analysis and Methods/Computational Optical Sensing and Imaging/Information Photonics/Signal Recovery and Synthesis Topical Meetings on CD-ROM, Technical Digest (OSA, 2005), paper JMA4.
14. D. Takhar, J. N. Laska, M. B. Wakin, M. F. Duarte, D. Baron, S. Sarvotham, K. Felly, and R. G. Baraniuk, “A new compressive imaging camera architecture using optical-domain compression,” in Proc. of Computational Imaging IVSPIE Electronic Imaging6065, San Jose, CA (2006).
15. H. S. Pal, D. Ganotra, and M. A. Neifeld, “Face recognition by using feature-specific imaging,” Appl. Opt. 44, 3784–3794 (2005). [CrossRef] [PubMed]
16. M. A. Neifeld and J. Ke, “Optical architectures for compressive imaging,” Appl. Opt. 46, 5293–5303 (2007). [CrossRef] [PubMed]
17. P. K. Baheti and M. A. Neifeld, “Feature-specific structured imaging,” Appl. Opt. 45, 7382–7391 (2006). [CrossRef] [PubMed]
18. E. J. Candès and J. Romberg “Practical signal recovery from random projections,” in Wavelet Applications in Signal and Image Processing XI, Proc. SPIE Conf.5914, (2004).
19. E. J. Candès, J. Romberg, and T. Tao “Stable signal recovery from incomplete and inaccurate measurements,” Comm. Pure Appl. Math. 59, 1207–1223 (2005). [CrossRef]
20. M. A. T. Figueiredo, R. D. Nowak, and S. J. Wright, “Gradient projection for sparse reconstruction: Application to compressed sensing and other inverse problems,” (Preprint, 2007). Available: http://www.lx.it.pt/~mtf/GPSR/
21. H. H. Barrett and K. J. Myers, Foundations of Image Science (Wiley series in Pure and Applied Optics, 2004).