Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Imaging through scattering media using speckle pattern classification based support vector regression

Open Access Open Access

Abstract

Imaging through scattering media is a common practice in many applications of biomedical imaging. Object image would deteriorate into unrecognizable speckle pattern when scattering media is presented. Many methods have been investigated to reconstruct the object image when only speckle pattern is available. In this paper, we demonstrate a method of single-shot imaging through scattering media. This method is based on classification and support vector regression of the measured speckle pattern. We prove the possibility of speckle pattern classification and related formulas are presented. The specified and limited imaging capability without speckle pattern classification is demonstrated. Our proposed approach, that is, speckle pattern classification based support vector regression method, makes up the deficiency. Experimental results show that, with our approach, speckle patterns could be utilized for classification when object images are unavailable, and object images can be reconstructed with high fidelity. The proposed approach for imaging through scattering media is expected to be applicable to various sensing schemes.

© 2018 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

Imaging or focusing through scattering media is a common practice in the fields of biomedicine and security [1–7]. Scattering media would diffuse the object image into unrecognizable speckle pattern. Various methods for retrieving the object image from scattered speckle pattern have been proposed. Wavefront shaping technique has emerged as a powerful tool for inverse scattering by controlling the wavefront of incident light [8–11]. However, these techniques are complex and lengthy, since a detector or an optical/acoustical probe is always necessary in the plane of interests to provide feed-back instructions for wavefront modulation. Phase-retrieval based method utilizes the principle of optical memory effects to translate the inverse scattering problem into phase retrieval problem [12–15]. But phase retrieval algorithms are always dependent on the initial points (so needs to restart several times to obtain a satisfied reconstruction) and noise sensitive (since noise would pollute the measured speckle pattern and introduce bias to Fourier amplitude of object image). Ghost imaging can retrieve the information of an unknown object without a spatial-resolving detector toward it, while a reference beam is necessary and sometimes a calibration is needed [16–19]. Machine-learning based sensing approach was introduced to realize imaging through scattering media without time-consuming feedback modulation and reference arm, and is insensitive to noise enough due to its data-driven character. However, the learned inverse scattering function (ISF) was only effective and specified for reconstruction of images located at the same category as training object-image-and-speckle-pattern pairs (OS pairs in short) [20–22]. Specifically, the ISF learned from OS pairs of class 1 was only able to reconstruct objects of class 1 from their speckle patterns well, while failed to reconstruct objects of other classes, and the reconstructions were always ambiguous profiles of objects of class 1.

Exactly the limited imaging capability of the learned ISF, inspired us to try to make classification to scattered speckle patterns [22]. In this paper, we propose to apply image classification to speckle patterns and then use speckle pattern classification based support vector regression (referenced as SPC-SVR) method for single-shot imaging through scattering media. Object classification and recognition has been a hot topic in optical remote sensing, biomedical imaging, and etc [23–26]. In these fields, the targets for classification or recognition are always the objects themselves. However, in a scattering system, objects would scratch into unrecognizable speckle patterns and become unavailable temporarily. The key for classification is to learn more discriminative features of images and determine the optimal label for the given image data. We will prove mathematically that the autocorrelation matrix of the scattered speckle pattern has the same singular values or matrix as that of the object image, and thus speckle patterns could be utilized for classification when object images are unavailable. To realize imaging through scattering media with our proposed SPC-SVR approach, first of all, a database containing adequate well-labeled known OS pairs should be established. The label of each speckle pattern in the known database is defined to be consistent with that of the corresponding object image (referenced as class 1, class 2 and etc.). Then, classification algorithms are used to learn classifiers with speckle patterns in the known OS Database. The considered classification algorithms are, two-dimensional principal component analysis and support vector machine (referenced as PCA+SVM), sparse representation based classification (SRC), K-singular value decomposition (K-SVD) and Kernel K-singular value decomposition (Kernel KSVD), while the considered regression model is support vector regression here [27–33]. Given an unknown speckle pattern, the learned classifiers are firstly adopted to classify the label of the speckle pattern, next, a certain number of OS pairs of the same label are randomly selected to learn an ISF, the learned ISF (of the same label) is then utilized to reconstruct the original object image. Our approach is attractive for several reasons: (1), we realize speckle pattern based classification without knowing the exact corresponding object images. (2), the classifiers, once learned, could be adopted to any unknown speckle patterns efficiently and there is no need to learn them each time. More examples per category would lead to higher classification accuracy. (3), the ISF, once learned through dataset with specific label, should reconstruct every possible object image with the same label efficiently. The more ISFs (of different labels) learned, the more objects (of different labels) would benefit from them and could be reconstructed effectively. Neither time-consuming calculations of scattering effect, nor separately iteration processes are needed for each new object under reconstruction. (4), our approach improves the aforementioned limited imaging capability, extending the practicability and feasibility of scattering imaging with machine-learning-based approach. The approach is expected to be applicable to image reconstruction in X-ray computed tomography, source reconstruction in acoustics, Earth density calculation, decryption in optics, and etc [34–37].

The remains of the paper are organized as follows : the methodology and the mathematical derivation of speckle pattern classification are given in Section 2. Then, we discuss the object image regression methodology based on support vector regression model in Section 3. Section 4 introduces the employed reconstruction fidelity evaluators. Section 5 demonstrates experimental results and quantitative analyses. Conclusions are drawn in Section 6.

2. Speckle pattern classification methodology

Our previous work showed that without classification first, the learned ISF with OS pairs of a certain class was only specified to reconstruct objects from speckles of that class [22]. To realize imaging through scattering media better, speckle pattern classification is proposed to conduct before object image reconstruction or regression. But in scattering imaging systems (see Fig. 1 for example), the object images are always scattered into unrecognizable speckle patterns. In this section, we present the deviation result that the autocorrelation matrix of the scattered speckle pattern has the same singular values or matrix as that of the object image, and thus could be utilized for classification. Since the key for classification is to learn more discriminative features of images so as to determine the optimal label for the given image data, the derivation is mainly based on singular value decomposition (SVD).

 figure: Fig. 1

Fig. 1 Experimental setup of a scattering system. O, objective; P, pinhole; L, lens; SLM, spatial light modulator.

Download Full Size | PDF

The simplified experimental arrangement of a single-layer scattering system is demonstrated in Fig. 1. Light from a He-Ne laser source is collimated and expanded, then modulated by the amplitude type spatial light modulator (SLM, HES6001, Holoeye, used to modulate object images, and the two polarizers surrounding the SLM are omitted), then the modulated light traveled through a diffuser (DG10-220-MD, Thorlabs, served as a scattering media), and the scattered light was captured by a CMOS camera (C13440-20CU, Hamamatsu). The distance between modulated object image and the diffuser is set to be 25cm, and the image distance from the diffuser to camera is 15cm. The relationship between input object image and corresponding output speckle pattern captured by the image sensor in the scattering system (see Fig. 1) can be described as:

Eoutejϕ=KEin,
where EoutRMout and ϕCMout are the vectorized amplitude field and phase field of output speckle pattern, respectively. ◦ denotes the Hadamard product operation. K represents the transmission matrix (TM) of scattering media (i.e., the diffuser here), with size Mout × Min. EinRMin represents the vectorized input object image. Min and Mout are the pixel numbers in each input object image and output speckle pattern, respectively. Note that, due to the limitation of image sensor, only the amplitude field of output speckle field can be accessed in experiment.

Multiply both sides of Eq. (1) by its conjugate transpose, then we obtain:

(Eoutejϕ)(Eoutejϕ)*=(KEin)(KEin)*=KEin(Ein)*K*,
where (Eoute) · (Eoute)* is a real symmetric matrix of size Mout × Mout. Similarly, Ein · (Ein) is a real symmetric matrix of size Min × Min. (·) means the conjugate transpose operation.

According to the transmission matrix theory, the TM K is a unitary matrix [3]. Then through performing SVD to Ein · (Ein)*, (Eoute) · (Eoute)* and Eout · (Eout)* separately, we draw the conclusion that the three matrices share the same singular values or singular value matrix (see appendix A for detailed derivations). If we arrange the singular values in an order from large to small by default, then the singular matrix (owned by the three aforementioned matrices) is deduced to be the autocorrelation of the singular matrix of Ein.

The conclusion that, Ein ·(Ein)* and Eout · (Eout)* own the same singular values, is meaningful and can be utilized for speckle classification when object is unavailable. Besides, the conclusion also indicates that speckle pattern classification should perform the same as object classification theoretically.

3. Object image regression methodology

Given an unknown speckle pattern, speckle pattern classification is firstly performed to determine an optimum label. Then based on a dataset of OS pairs with the same label, support vector regression (SVR) is utilized to learn a corresponding ISF. The methodology of the utilized SVR model is introduced as follows.

The relationship between output speckle pattern Eout and input object image Ein can be rewritten as:

Eout=f(Ein),Ein=f1(Eout),
where f (·) and f−1(·) denotes the forward scattering function (SF) and ISF, respectively.

Inverse scattering is to retrieve the object image from its speckle pattern. The flow chart of our approach is showed in Fig. 2. Given an unknown speckle pattern, the first thing to do is speckle pattern classification based on pre-established database (of relationships between known speckle patterns and their labels, SL in short). The flow chart of speckle pattern classification is showed in Fig. 3. Assume that the unknown speckle pattern is classified and labeled with P, then the corresponding ISF would be learned with that class of OS pairs, to conduct regression to reconstruct the object image in a further step.

 figure: Fig. 2

Fig. 2 Flow chart of the inverse scattering method.

Download Full Size | PDF

 figure: Fig. 3

Fig. 3 Flow chart of speckle pattern classification.

Download Full Size | PDF

The ISF learning with certain known OS pairs of a certain class is modelled by solving the following support vector regression problem:

minw,b12wTw+Cn=1Nmax(0,|wTEnout+bEnin|ε),
where w is the inverse sensing matrix and follows f−1(Eout) = wTEout + b (b is the intercept vector of ISF), (·)T means the transpose operation, C means a constant parameter trading between regularization and violation, Enout means the nth output speckle, Enin represents the nth input object, ε represents a parameter indicating the acceptable error, N is the number of training pairs. For each OS pair, training is conducted pixel by pixel.

As for any L2-regularized linear model, their optimal solution can be represented as the linear combination of independent variables, that is, w=n=1NβnEnout, where βn is the coefficient corresponding to nth speckle pattern. Substituting it to Eq. (4) and applying the kernel trick (where the radial basis function is served as kernel), we obtain:

minβ,b12n=1Nβnβmexp(γ||EnoutEmout||2)+Cn=1Nmax(0,|m=1Nβmexp(γ||EnoutEmout||2)Enin|ε),
where exp(γ||EnoutEmout||2) denotes the applied radial basis kernel function (RBF) in regression process. Solving the problem described in Eq. (5), the ISF can be learned and then the object can be reconstructed finally.

With the methodologies described above, unknown speckle patterns can be classified and the object situated behind scattering media can be reconstructed effectively. The next section introduces the employed indicators for image reconstruction fidelity evaluation.

4. Image reconstruction fidelity evaluation

To evaluate the quality of the scattering image reconstruction results objectively, the peak signal-to-noise ratio (PSNR) and the structural similarity (SSIM) are used in this paper [38, 39]. Both are commonly and widely used objective evaluation standard of image quality. The PSNR between image x and image y of same size px × py is defined as:

PSNR(dB)=10log10(MAXI 2MSE),MSE=1pxpyi=0px1j=0py1[x(i,j)y(i,j)]2,
where MAXI denotes the maximum possible pixel value of images and MAXI = 1 since a normalization process is conducted before classification and regression. px and py represent the number of pixels of images in x-axis and y-axis, respectively.

The SSIM between image x and image y is defined as

SSIM(x,y)=(2μxμy+c1)(2σxy+c2)(μx2+μy2+c1)(σx2+σy2+c2),
where µx and µy are the mean value of image x and y respectively, σx and σy are the variance of image x and y respectively, σxy is the covariance of image x and y. c1 and c2 are small positive constants used to avoid a null denominator. SSIM is a decimal value between −1 and 1. When the reconstructed image is identical to our input object image (which means that perfect reconstruction is realized), the SSIM reaches value 1.

Substituting image x and image y with our input object image and the reconstructed one respectively, then the PSNR value, as well as the SSIM value between them, can be calculated.

Next, experimental results are demonstrated to verify the methodologies illustrated above.

5. Experimental results

5.1. Database establishment

Before inverse scattering, a database containing well-labeled known OS pairs should be established. To reflect the importance of classification before regression well, in the paper, the MNIST handwritten digit dataset (referenced as MNIST) and the Fashion MNIST dataset (a collection of Zalando’s article images, referenced as Fashion MNIST) are utilized to provide object images [38, 40]. The considered two datasets are all extensively used in image classification, optical character recognition and machine learning research, sharing the same image size of 28 28 and structure of training and testing splits. The MNIST provides 60, 000 training examples×and 10, 000 testing examples, and the distributions of digits over different classes are detailed in Table. 1. The Fashion MNIST is also composed of a training set of 60, 000 examples and a testing set of 10, 000 examples, associating with a label from 10 classes. For each class, there is 6, 000 training examples and 1, 000 testing examples in the Fashion MNIST. The MNIST consisting of a collection of digit images (from digit “0” to “9”, handwritten in different styles), while images in Fashion MNIST are ranging from “T-shirt” to “Ankle boot”. For the convenient of learning the speckle classifiers, as well as the ISFs for speckle of certain label, images are all enlarged to a same size to serve as the input objects of the scattering system (see Fig. 1) one by one, and the corresponding output speckles are also collected one by one. As stated before, each output speckle is labeled according to the label of its input object image. Besides, normalization preprocessing was considered for all object images. In that case, our MNIST OS Dataset and Fashion MNIST OS Dataset can be established, respectively. Figure. 4 and Fig. 5 gives the illustration of some examples (of OS pairs) from the MNIST OS Dataset and the Fashion MNIST OS Dataset.

Tables Icon

Table 1. The example numbers in the MNIST Database.

 figure: Fig. 4

Fig. 4 Illustrations of different OS pairs from the MNIST OS Dataset. From left to right, each column shows normalized OS pair of a certain class from “0” to “4”. (a) object image of “0”, (b) speckle pattern of (a); (c) object image of “1”, (d) speckle pattern of (c); (e) object image of “2”, (f) speckle pattern of (e); (g) object image of “3”, (h) speckle pattern of (g); (i) object image of “4”, (j) speckle pattern of (i).

Download Full Size | PDF

 figure: Fig. 5

Fig. 5 Illustrations of different OS pairs from the Fashion MNIST OS Dataset. From left to right, each column shows normalized OS pair of a certain class from “T-shirt” to “Sneaker”. (a) object image of “T-shirt”, (b) speckle pattern of (a); (c) object image of “Trouser”, (d) speckle pattern of (c); (e) object image of “Coat”, (f) speckle pattern of (e); (g) object image of “Sandals”, (h) speckle pattern of (g); (i) object image of “Sneaker”, (j) speckle pattern of (i).

Download Full Size | PDF

5.2. limited imaging capability without classification first

Our previous work showed that, without classification before regression or reconstruction, the capability of imaging through scattering media of machine-learning-based method is limited [22]. The limited imaging capability has also been discussed before, where the ISF learned from human faces was only specified for reconstructing human face from their speckle patterns [21]. In a word, with the ISF trained from OS pairs of class 1, the machine-learning-based regression method could only be able to reconstruct objects (of the same class) from their unknown speckles. The limitation is also held up for the ISF trained from OS pairs of any other classes.

Experiments were conducted to demonstrate the limited imaging capability reflected on the MNIST OS Dataset and Fashion MNIST OS Dataset, respectively. Both object images and their speckle patterns are all sampled to 28 × 28, i.e., Min = 784. To ensure adequately learning, the number of training OS pairs is set as 784 (i.e., Mout = 784). The OS pairs used for learning an ISF of a certain label, are randomly chosen from the exact subset of the OS Database without overlap. In Fig. 6 and Fig. 7, each row is an example, where the listed images in the first four columns are the normalized input object image, measured speckle pattern, reconstruction with ISF1, reconstruction with ISF2 successively from left to right. As for each reconstruction, the calculated fidelity evaluation indicators (i.e., the PSNR and SSIM values) were summarized and inserted at the bottom of the corresponding subfigure. Combining the reconstructions and the indicators, the SSIM indicator seems to be relatively more objective. In Fig. 6, the utilized ISF in the third column is learned from digit “0” OS pairs, while that in the forth column is learned from digit “4” OS pairs. In Fig. 7, the utilized ISF in the third column is learned from “Trouser” OS pairs, while that in forth column is learned from “Sneaker” OS pairs.

 figure: Fig. 6

Fig. 6 Reconstructions without classification first before learning the ISF with the MNIST OS Dataset. From left to right, the listed images in each column are, (a) and (b) input object images, (c) and (d) speckle patterns, (e) and (f) reconstructions with ISF learned from digit “O” OS pairs, (g) and (h) reconstructions with ISF learned from digit “4” OS pairs, (i) and (j) reconstructions with ISF learned from OS pairs of all the 10 classes.

Download Full Size | PDF

 figure: Fig. 7

Fig. 7 Reconstructions without classification first before learning the ISF with the Fashion MNIST OS Dataset. From left to right, the listed images in each column are, (a) and (b) input object images, (c) and (d) speckle patterns, (e) and (f) reconstructions with ISF learned from “Trouser” OS pairs, (g) and (h) reconstructions with ISF learned from “Sneaker” OS pairs, (i) and (j) reconstructions with ISF learned from OS pairs of all the 10 classes.

Download Full Size | PDF

Experiments conducted with the MNIST OS Dataset and the Fashion MNIST OS Dataset showed that, the ISF learned with OS pairs of a class can regress or reconstruct objects of the class well, but the imaging capability of the ISF was restricted to objects of the class, and the ISF failed to reconstruct objects of other classes. The restriction was much severer in Fig. 7. This is due to the bigger difference among classes in the Fashion MNIST, which can also be reflected from the relatively lower classification accuracy in next subsection [40]. The phenomenon of limited imaging capability is consistent with that in our previous work [22]. The ISF learned from OS pairs of a certain class always has its limitation to reconstruct images of the other classes from their speckle patterns.

Including more classes of OS pairs to learn an ISF may help. Experiments were conducted to verify. In the experiments, the training OS pairs were randomly but about averagely chosen from all the 10 classes. The other experimental conditions were set the same as former. For comparison, results are listed in the fifth column of Fig. 6 and Fig. 7, respectively. Results showed that, with training OS pairs of more classes, even the learned ISF could regress speckles of the other class to some extent, but the image reconstruction fidelities are relatively lower than those whose speckles were regressed with ISF of the same class. The more ambiguous background should be responsible for this.

To improve the situation, speckle pattern classification is proposed. Next, we’ll show that, with speckle pattern classification based support vector regression method, capability and fidelity of imaging through scattering media can be improved at high probability.

5.3. imaging with speckle pattern classification based support vector regression

Experiments were conducted to validate speckle pattern classification performance. For each experiment, a certain number of OS pairs were chosen from training set of a dataset (i.e., the MNIST OS Dataset or the Fashion MNIST OS Dataset here) for learning classifiers, while speckle patterns (from testing set of the Datasets) were served as test or unknown ones to examine the performances of the classifiers learned through some related classification methods, then recorded the corresponding accuracies. Speckle pattern classification accuracies with several algorithms, i.e., PCA+SVM, SRC, K-SVD and Kernel KSVD, were considered here. Parameters of these algorithms were all empirically determined. The used kernel for SVM was gaussian radial basis function whose parameter γ was set as 0.25, and the number of feature vectors were extracted to 20. In SRC, the sparsity level was set as 20, all training samples were used to generate an overcomplete dictionary. In K-SVD and Kernel KSVD, dictionaries were learned with 30 atoms, the sparsity level was set as 20, and the maximum number of training iterations was set as 60. For kernel KSVD, a polynomial kernel of degree 4 was used. Classification accuracies under different number of training samples were also considered. For comparison, object classification was also conducted. The training samples were objects in object classification and turned to speckle patterns in speckle pattern classification. Table 2 and Table 3 shows the classification accuracies on the MNIST OS Database and the Fashion MNIST OS Database, respectively. Po is the probability of correct object classification and Ps is the probability of correct speckle classification.

Tables Icon

Table 2. Speckle pattern classification accuracies of some related methods on the MNIST OS Database.

Tables Icon

Table 3. Speckle pattern classification accuracies of some related methods on the Fashion MNIST OS Database.

From the speckle pattern classification results listed in Table 2 and Table 3, several points could be concluded. First, the seemly unrecognizable speckle patterns could actually be used for image classification indeed and the satisfying speckle pattern classification accuracies were closed to the object classification accuracies, which validated the methodology described before. Second, with more training samples included for classifier learning, both the object and speckle pattern classification accuracies got some improvements for all considered algorithms. Besides, the kernel KSVD always performs best among all the considered algorithms.

Considering our proposed SPC-SVR method, higher classification accuracy means better reconstruction fidelity at higher probability. Once we get the label of an unknown speckle pattern, the exact ISF (learned from OS pairs of the same label) can be applied for object reconstruction. In the experiment, all OS pairs in training set were used for learning classifiers, while 784 (= 28 × 28) OS pairs were randomly chosen from training set for learning ISF of a certain label. Figure. 8 and Fig. 9 demonstrated the object reconstruction results based on the MNIST OS Dataset and the Fashion MNIST OS Dataset, respectively. The corresponding reconstruction fidelity evaluators were summarized and inserted at the bottom of reconstruction results. As shown in the figures, the proposed approach realized high fidelity imaging through scattering media at high probability based on the MNIST dataset and Fashion MNIST dataset. The resulting averaged PSNR value was about 15.90dB and the averaged SSIM value was about 0.42 based on the MNIST dataset, while the averaged PSNR value was about 20.67dB and the averaged SSIM value was about 0.65 based on the Fashion MNIST dataset. The difference was mainly due to the different sparsity levels of samples in the two datasets. The results were consistent with those in Fig. 6 and Fig. 7. Besides, the reconstruction or regression for an object took about one minute where the time for ISF learning was not included.

 figure: Fig. 8

Fig. 8 Reconstructions using speckle pattern classification based support vector regression with the MNIST OS Database. (a) ∼ (j) show object and reconstruction examples of all the 10 classes from digit “0” to “9”.

Download Full Size | PDF

 figure: Fig. 9

Fig. 9 Reconstructions using speckle pattern classification based support vector regression with the Fashion MNIST OS Database. (a) ∼ (j) show object and reconstruction examples of all the 10 classes from “T-shirt” to “Ankle boots”.

Download Full Size | PDF

5.4. Comparisons

Comparisons were conducted to validate effectiveness of our proposed method. Reconstruction results of wavefront shaping technique, ghost imaging, and phase-retrieval based method are also demonstrated. For our method, after speckle pattern classification finished, 784 known OS pairs were utilized to learn an ISF for reconstruction. For an arbitrary given unknown speckle pattern, the averaged time for classification was about 5 seconds using Kernel KSVD and that for regression was about 144 seconds. For wavefront shaping technique, genetic algorithm was used to seek for a global optimized modulation on phase type SLM and the imaging time was about 1 hour for 1000 generations. For ghost imaging technique, 10,000 acquisitions of scattered light field and reference light field were used for reconstruction. For phase-retrieval based method, 1000 iterations are adequate for convergency. The reconstructions of compared methods tested on the MNIST Database and the Fashion MNIST Database were showed in Fig. 10 and Fig. 11, respectively.

 figure: Fig. 10

Fig. 10 Reconstructions based on MNIST Database with different methods. (a) tested original object image; (b) reconstruction with the proposed method; (c) reconstruction with wavefront shaping technique; (d) reconstruction with ghost imaging technique; (e) reconstruction with phase-retrieval based method.

Download Full Size | PDF

 figure: Fig. 11

Fig. 11 Reconstructions based on Fashion MNIST Database with different methods. (a) tested original object image; (b) reconstruction with the proposed method; (c) reconstruction with wavefront shaping technique; (d) reconstruction with ghost imaging technique; (e) reconstruction with phase-retrieval based method.

Download Full Size | PDF

The phase-retrieval based method seemed to be the most convenient one, since reconstruction can be realized with just a single-shot measurement. Our proposed method required a well-labeled dataset to establish a Database containing adequate OS pairs. The wavefront shaping technique realized reconstruction through time-consuming feedback modulations. And the ghost imaging technique recorded the corresponding scattered light field and reference light field when random phase pattern was modulated on SLM one by one, to calculate image of object. From reconstructions in Fig. 10 and Fig. 11, where ghost imaging technique suffered, one could conclude that reconstruction fidelity of the proposed method was the highest.

6. Conclusions

In this paper, we deduced that in a scattering system, the autocorrelation of an input object image and the autocorrelation of the output speckle pattern have the same singular values, and propose that the seemly unrecognizable speckle patterns can be utilized for image classification when objects are unavailable. Speckle pattern classification based support vector regression method is proposed and utilized for single-shot imaging through scattering media. Speckle pattern classification is used to obtain the label of a given unknown speckle pattern, while support vector regression is designed for learning inverse scattering function of the scattering system, as well as original object reconstruction or regression. Experiments show that classifier with higher speckle pattern classification accuracy means higher object reconstruction or regression fidelity. Comparisons with some existing methods also validate the effectiveness of the proposed method. The proposed speckle pattern classification based support vector regression method, not only makes up the limitation of our previous research, but also provides a generalized and convenient approach for imaging through scattering media. Besides, our method could actually be seen as a generalized and universal solution for inverse problems such as phase retrieval and etc.

Our approach performs well in our setup, but a sufficiently large database of known object images and the corresponding speckle patterns should be established before learning the classifier and ISF for each category. More examples of speckle patterns mean a more stable and knowledgeable classifier, as well as higher reconstruction fidelity or more capable imaging through scattering media. Future issues to be addressed include, developing a efficient and efficiency image classification method (such as the convolutional neural nets) to ensure high speckle pattern classification accuracy and satisfying image reconstruction fidelity, multidimensional object reconstruction with the speckle pattern classification based support vector regression.

Appendix A - detailed derivations about the principle of speckle classification

In this appendix, we give the detailed derivations of the conclusion that Ein · (Ein), (E oute) (E oute) and Eout · (Eout)* own the same singular values.

According to the transmission matrix theory, the TM K is a unitary matrix [3]. That is, K · K = I (I is identity matrix). The singular value decomposition (SVD) of K can be defined as:

K=S1V1D1*,
where S1 and D1 are unitary matrices, following S1S1*=S1*S1=I and D1D1*=D1*D1=I respectively, V1 (whose diagonal elements are arranged in an order from large to small) is the singular value matrix of K. For convenience, the diagonal elements of singular value matrices mentioned in the paper are all arranged in an order from large to small.

Substituting Eq. (8) and K · K = I, and we obtain:

KK*=(S1V1D1*)(S1V1D1*)*=S1V1D1*D1V1*S1*=S1V1V1*S1*=I.

Meanwhile, define the SVD of Ein as:

Ein=S2V2D2*,
where S2 and D2 are unitary matrices, V2 is the singular value matrix of Ein.

Substituting Eq. (8) to Eq. (10) into (Eoute) · (Eoute)* = K · Ein · (Ein)* · K* (i.e., the Eq. (2) in the main body), we get:

(Eoutejϕ)(Eoutejϕ)*=KEin(Ein)*K*=(S1V1D1*)(S2V2D2*)(S2V2D2*)*(S1V1D1*)*=S1V1D1*S2V2D2*D2V2*S2*D1V1*S1*=S1V1D1*S2V2V2*S2*D1V1*S1*=(S1V1D1*S2)(V2V2*)(S1V1D1*S2)*.

Obviously, V2V2* is still a diagonal matrix where each element equals to the square that in V2. The necessary The necessary and sufficient condition for that V2V2* is the singular value matrix of (Eoute) · (Eoute)* is that S1V1D1*S2 is a unitary matrix.

Then the problem is converted to prove S1V1D1*S2 (denoted as S3 for short) to be a unitary matrix.

Multiply S3 by its Hermitian transpose, and substitute Eq. (9) into the product, then we obtain:

S3(S3)*=(S1V1D1*S2)(S1V1D1*S2)*=S1V1D1*S2S2*D1V1*S1*=S1V1D1*D1V1*S1*=S1V1V1*S1*=I.

In Eq. (12), we have proven that S3 is also a unitary matrix, and thus Eq. (11) can be rewritten as:

(Eoutejϕ)(Eoutejϕ)*=S3V3D3*,
where S3=D3=S1V1D1*S2, V3=V2V2* is the singular value matrix of (Eoute) · (Eoute)*.

Return back to Eq. (10), the SVD of Ein · (Ein)* can be written as:

Ein(Ein)*=(S2V2D2*)(S2V2D2*)*=S2V2D2*D2V2*S2*=S2(V2V2*)S2*.

That is, V2V2* is also the singular value matrx of Ein · (Ein)*.

At this point, we have proven that, (Eoute) · (Eoute)* and Ein · (Ein)* shares the same values singular value matrix. However, only the amplitude field of the speckle pattern (i.e., Eout) could be accessible in experiment. In a word, it’s necessary to find some relationship between the singular value matrix of Eout · (Eout) and (Eoute) · (Eoute)*.

We’ll prove that Eout · (Eout)* has exactly the same singular value matrix as (Eoute) · (Eoute)*, as well as Ein · (Ein)*.

Reformat the vector ejϕ=[ejϕ1,ejϕ2,,ejϕMout]T (where the (·)T denotes transport operation) as a diagonal matrix (marked with φ), where each element in the diagonal is a corresponding element in e, i.e.,

ejϕ=[ejϕ1,ejϕ2,,ejϕMout]T[ejϕ1    ejϕ2        ejϕMout]=φ.

Then we have:

(Eoutejϕ)(Eoutejϕ)*=(φEout)(φEout)*=φEout(Eout)*φ*.

Define the SVD of Eout as:

Eout=S4V4D4*,
where S4 and D4 are unitary matrices, V4 is the singular value matrix of Eout.

Then we have:

Eout(Eout)*=(S4V4D4*)(S4V4D4*)*=S4V4D4*D4V4*S4*=S4(V4V4*)S4*.

Apparently, V4V4* is a diagonal matrix, and is the singular value matrix of Eout · (Eout). Substitute Eq. (18) to Eq. (16), then we obtain:

(Eoutejϕ)(Eoutejϕ)*=φEout(Eout)*φ*=φS4V4D4*(S4V4D4*)*φ*=φS4V4D4*D4V4*S4*φ*=(φS4)(V4V4*)(φS4)*.

It’s obviously that φS4 is a unitary matrix since (φS4)(φS4)*=φS4S4*φ*=φφ*=I. That is to say, V4V4* is the singular value matrix of (Eoute) · (Eoute)*. Besides, considering the uniqueness of the singular value matrices (given the defination that the singular values in the singular value matrix displayed here are all arranged in an order from large to small), we have V4V4*=V2V2* At this point, we can draw a conclusion that Ein · (Ein), (Eoute) · (Eoute)* and Eout · (Eout)*, share the same singular values or singular value matrix.

The conclusion can then serve as the principle of speckle pattern classification, that is, speckle patterns can be utilized for image classification and should have almost the same performance as object classification.

Funding

National Natural Science Foundation of China (Number:61601285).

Acknowledgments

The authors want to express their gratitude to editors and anonymous reviewers who gave their valuable comments and suggestions to this article.

Disclosures

The authors declare that there are no conflicts of interest related to this article.

References

1. Y. Wang, N. Cao, Z. Liu, and Y. Zhang, “Real-time dynamic MRI using parallel dictionary learning and dynamic total variation,” Neurocomputing 238, 410–419 (2017). [CrossRef]  

2. R. Horstmeyer, H. Ruan, and C. Yang, “Guidestar-assisted wavefront-shaping methods for focusing light into biological tissue,” Nat. Photonics 9, 563 (2015). [CrossRef]   [PubMed]  

3. M. Kim, W. Choi, Y. Choi, C. Yoon, and W. Choi, “Transmission matrix of a scattering medium and its applications in biophotonics,” Opt. Express 23, 12648–12668 (2015). [CrossRef]   [PubMed]  

4. J. Jang, J. Lim, H. Yu, H. Choi, J. Ha, J. H. Park, W. Y. Oh, W. Jang, S. Lee, and Y. Park, “Complex wavefront shaping for optimal depth-selective focusing in optical coherence tomography,” Opt. Express 21, 2890–2902 (2013). [CrossRef]   [PubMed]  

5. B. R. Anderson, P. Price, R. Gunawidjaja, and H. Eilers, “Microgenetic optimization algorithm for optimal wavefront shaping,” Appl. Opt. 54, 1485–1491 (2015). [CrossRef]   [PubMed]  

6. L. Fang, X. Zhang, H. Zuo, and L. Pang, “Focusing light through random scattering media by four-element division algorithm,” Opt. Commun. 407, 301–310 (2018). [CrossRef]  

7. L. Zhou, B. Zhuang, H. Sun, Z. He, M. Hu, and X. Qiao, “Speckle phase retrieval and transmission matrix obtaining of turbid media,” Opt. Int. J. Light Electron. Opt. 127, 9911–9916 (2016). [CrossRef]  

8. I. M. Vellekoop and A. Mosk, “Focusing coherent light through opaque strongly scattering media,” Opt. Lett. 32, 2309–2311 (2007). [CrossRef]   [PubMed]  

9. A. P. Mosk, A. Lagendijk, G. Lerosey, and M. Fink, “Controlling waves in space and time for imaging and focusing in complex media,” Nat. Photonics 6, 283 (2012). [CrossRef]  

10. I. M. Vellekoop, A. Lagendijk, and A. P. Mosk, “Exploiting disorder for perfect focusing,” Nat. Photonics 4, 320 (2010). [CrossRef]  

11. H. He, Y. Guan, and J. Zhou, “Image restoration through thin turbid layers by correlation with a known object,” Opt. Express 21, 12539–12545 (2013). [CrossRef]   [PubMed]  

12. O. Katz, P. Heidmann, M. Fink, and S. Gigan, “Non-invasive single-shot imaging through scattering layers and around corners via speckle correlations,” Nat. Photonics 8, 784 (2014). [CrossRef]  

13. E. Edrei and G. Scarcelli, “Optical imaging through dynamic turbid media using the Fourier-domain shower-curtain effect,” Optica 3, 71–74 (2016). [CrossRef]   [PubMed]  

14. P. Wu, Z. Liang, X. Zhao, L. Su, and L. Song, “Lensless wide-field single-shot imaging through turbid media based on object-modulated speckles,” Appl. Opt. 56, 3335–3341 (2017). [CrossRef]   [PubMed]  

15. Y. Shechtman, Y. C. Eldar, O. Cohen, H. N. Chapman, J. Miao, and M. Segev, “Phase retrieval with application to optical imaging: a contemporary overview,” IEEE Signal Process. Mag. 32, 87–109 (2015). [CrossRef]  

16. X.-H. Chen, Q. Liu, K.-H. Luo, and L.-A. Wu, “Lensless ghost imaging with true thermal light,” Opt. Lett. 34, 695–697 (2009). [CrossRef]   [PubMed]  

17. R. E. Meyers, K. S. Deacon, and Y. Shih, “Turbulence-free ghost imaging,” Appl. Phys. Lett. 98, 041801 (2011). [CrossRef]  

18. Y. Xu, W. Liu, E. Zhang, Q. Li, H. Dai, and P. Chen, “Is ghost imaging intrinsically more powerful against scattering?” Opt. Express 23, 32993–33000 (2015). [CrossRef]  

19. A. Zhang, Y. He, L. Wu, L. Chen, and B. Wang, “Tabletop x-ray ghost imaging with ultra-low radiation,” Optica 5, 374–377 (2018). [CrossRef]  

20. R. Horisaki, R. Takagi, and J. Tanida, “Learning-based focusing through scattering media,” Appl. Opt. 56, 4358–4362 (2017). [CrossRef]   [PubMed]  

21. R. Horisaki, R. Takagi, and J. Tanida, “Learning-based imaging through scattering media,” Opt. Express 24, 13738–13743 (2016). [CrossRef]   [PubMed]  

22. H. Chen, Y. Gao, and X. Liu, “Speckle reconstruction method based on machine learning,” Proc. SPIE 10711, 107111U (2018).

23. Q. Wang, L. Ma, C. Li, Y. Zhou, and L. Tang, “Fast object recognition method from random measurements of compressive sensing camera,” Proc. SPIE 10679, 1067919 (2018).

24. Y. Cheng, Y. Cao, Q. Zhang, Z. Chu, and R. K. Wang, “Deep network for retinal disease classification based on limited clinical OCT angiography datasets (Conference Presentation),” Proc. SPIE 10474, 1047407 (2018).

25. G. Ran, M. F. Stuhlmacher, B. Tellman, N. Clinton, G. Hanson, M. Georgescu, C. Wang, F. Serrano-Candela, A. K. Khandelwal, and W. H. Cheng, “Using landsat and nighttime lights for supervised pixel-based image classification of urban land cover,” Remote Sens. Environ. 205, 253–275 (2018). [CrossRef]  

26. S. H. Wang, P. Phillips, Z. C. Dong, and Y. D. Zhang, “Intelligent Facial Emotion Recognition based on Stationary Wavelet Entropy and Jaya algorithm,” Neurocomputing 272, 668–676 (2018). [CrossRef]  

27. M. O. Faruqe and M. A. M. Hasan, “Face recognition using PCA and SVM,” in International Conference on Anti-Counterfeiting, Security, and Identification in Communication (2009), pp. 97–101.

28. J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, “Robust Face Recognition via Sparse Representation,” IEEE Trans. Pattern Anal. Mach. Intell. 31, 210–227 (2009). [CrossRef]  

29. M. Aharon, M. Elad, and A. Bruckstein, “K-SVD: An Algorithm for Designing Overcomplete Dictionaries for Sparse Representation,” IEEE Trans. Signal Process. 54, 4311–4322 (2006). [CrossRef]  

30. H. V. Nguyen, V. M. Patel, N. M. Nasrabadi, and R. Chellappa, “Kernel dictionary learning,” in IEEE International Conference on Acoustics, Speech and Signal Processing, (2012), pp. 2021–2024.

31. H. Van Nguyen, V. M. Patel, N. M. Nasrabadi, and R. Chellappa, “Design of non-linear kernel dictionaries for object recognition,” IEEE Trans. Image Process. 22, 5123–5135 (2013). [CrossRef]   [PubMed]  

32. Z. Wang, Y. Wang, H. Liu, and H. Zhang, “Structured Kernel Dictionary Learning with Correlation Constraint for Object Recognition,” IEEE Trans. Image Process. 26, 4578–4590 (2017). [CrossRef]  

33. Vapnik and N. Vladimir, “The nature of statistical learning theory,” IEEE Trans. Neural Netw. 8, 1564 (1997). [CrossRef]  

34. Y. Cai, D. Liu, J. P. Mathews, Z. Pan, D. Elsworth, Y. Yao, J. Li, and X. Guo, “Permeability evolution in fractured coal−combining triaxial confinement with X-ray computed tomography, acoustic emission and ultrasonic techniques,” Int. J. Coal Geol. 122, 91–104 (2014). [CrossRef]  

35. Y. Kim and P. Nelson, “Optimal regularisation for acoustic source reconstruction by inverse methods,” J. Sound Vib. 275, 463–487 (2004). [CrossRef]  

36. J. Wahr, M. Molenaar, and F. Bryan, “Time variability of the Earth’s gravity field: Hydrological and oceanic effects and their possible detection using GRACE,” J. Geophys. Res. Solid Earth 103, 30205–30229 (1998). [CrossRef]  

37. M. Joshi, Chandrashakher, and K. Singh, “Color image encryption and decryption for twin images in fractional Fourier domain,” Opt. Commun. 281, 5713–5720 (2008). [CrossRef]  

38. Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proc. IEEE 86, 2278–2324 (1998). [CrossRef]  

39. Z. Wang, E. P. Simoncelli, and A. C. Bovik, “Multi-scale structural similarity for image quality assessment,” in Asilomar Conference on Signals Systems and Computers (2003), pp. 1398–1402.

40. H. Xiao, K. Rasul, and R. Vollgraf, “Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms,” (2017).

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (11)

Fig. 1
Fig. 1 Experimental setup of a scattering system. O, objective; P, pinhole; L, lens; SLM, spatial light modulator.
Fig. 2
Fig. 2 Flow chart of the inverse scattering method.
Fig. 3
Fig. 3 Flow chart of speckle pattern classification.
Fig. 4
Fig. 4 Illustrations of different OS pairs from the MNIST OS Dataset. From left to right, each column shows normalized OS pair of a certain class from “0” to “4”. (a) object image of “0”, (b) speckle pattern of (a); (c) object image of “1”, (d) speckle pattern of (c); (e) object image of “2”, (f) speckle pattern of (e); (g) object image of “3”, (h) speckle pattern of (g); (i) object image of “4”, (j) speckle pattern of (i).
Fig. 5
Fig. 5 Illustrations of different OS pairs from the Fashion MNIST OS Dataset. From left to right, each column shows normalized OS pair of a certain class from “T-shirt” to “Sneaker”. (a) object image of “T-shirt”, (b) speckle pattern of (a); (c) object image of “Trouser”, (d) speckle pattern of (c); (e) object image of “Coat”, (f) speckle pattern of (e); (g) object image of “Sandals”, (h) speckle pattern of (g); (i) object image of “Sneaker”, (j) speckle pattern of (i).
Fig. 6
Fig. 6 Reconstructions without classification first before learning the ISF with the MNIST OS Dataset. From left to right, the listed images in each column are, (a) and (b) input object images, (c) and (d) speckle patterns, (e) and (f) reconstructions with ISF learned from digit “O” OS pairs, (g) and (h) reconstructions with ISF learned from digit “4” OS pairs, (i) and (j) reconstructions with ISF learned from OS pairs of all the 10 classes.
Fig. 7
Fig. 7 Reconstructions without classification first before learning the ISF with the Fashion MNIST OS Dataset. From left to right, the listed images in each column are, (a) and (b) input object images, (c) and (d) speckle patterns, (e) and (f) reconstructions with ISF learned from “Trouser” OS pairs, (g) and (h) reconstructions with ISF learned from “Sneaker” OS pairs, (i) and (j) reconstructions with ISF learned from OS pairs of all the 10 classes.
Fig. 8
Fig. 8 Reconstructions using speckle pattern classification based support vector regression with the MNIST OS Database. (a) ∼ (j) show object and reconstruction examples of all the 10 classes from digit “0” to “9”.
Fig. 9
Fig. 9 Reconstructions using speckle pattern classification based support vector regression with the Fashion MNIST OS Database. (a) ∼ (j) show object and reconstruction examples of all the 10 classes from “T-shirt” to “Ankle boots”.
Fig. 10
Fig. 10 Reconstructions based on MNIST Database with different methods. (a) tested original object image; (b) reconstruction with the proposed method; (c) reconstruction with wavefront shaping technique; (d) reconstruction with ghost imaging technique; (e) reconstruction with phase-retrieval based method.
Fig. 11
Fig. 11 Reconstructions based on Fashion MNIST Database with different methods. (a) tested original object image; (b) reconstruction with the proposed method; (c) reconstruction with wavefront shaping technique; (d) reconstruction with ghost imaging technique; (e) reconstruction with phase-retrieval based method.

Tables (3)

Tables Icon

Table 1 The example numbers in the MNIST Database.

Tables Icon

Table 2 Speckle pattern classification accuracies of some related methods on the MNIST OS Database.

Tables Icon

Table 3 Speckle pattern classification accuracies of some related methods on the Fashion MNIST OS Database.

Equations (19)

Equations on this page are rendered with MathJax. Learn more.

E o u t e j ϕ = K E i n ,
( E o u t e j ϕ ) ( E o u t e j ϕ ) * = ( K E i n ) ( K E i n ) * = K E i n ( E i n ) * K * ,
E o u t = f ( E i n ) , E i n = f 1 ( E o u t ) ,
min w , b 1 2 w T w + C n = 1 N max ( 0 , | w T E n o u t + b E n i n | ε ) ,
min β , b 1 2 n = 1 N β n β m exp ( γ | | E n o u t E m o u t | | 2 ) + C n = 1 N max ( 0 , | m = 1 N β m exp ( γ | | E n o u t E m o u t | | 2 ) E n i n | ε ) ,
P S N R ( d B ) = 10 log 10 ( M A X I   2 M S E ) , M S E = 1 p x p y i = 0 p x 1 j = 0 p y 1 [ x ( i , j ) y ( i , j ) ] 2 ,
S S I M ( x , y ) = ( 2 μ x μ y + c 1 ) ( 2 σ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 2 ) ,
K = S 1 V 1 D 1 * ,
K K * = ( S 1 V 1 D 1 * ) ( S 1 V 1 D 1 * ) * = S 1 V 1 D 1 * D 1 V 1 * S 1 * = S 1 V 1 V 1 * S 1 * = I .
E i n = S 2 V 2 D 2 * ,
( E o u t e j ϕ ) ( E o u t e j ϕ ) * = K E i n ( E i n ) * K * = ( S 1 V 1 D 1 * ) ( S 2 V 2 D 2 * ) ( S 2 V 2 D 2 * ) * ( S 1 V 1 D 1 * ) * = S 1 V 1 D 1 * S 2 V 2 D 2 * D 2 V 2 * S 2 * D 1 V 1 * S 1 * = S 1 V 1 D 1 * S 2 V 2 V 2 * S 2 * D 1 V 1 * S 1 * = ( S 1 V 1 D 1 * S 2 ) ( V 2 V 2 * ) ( S 1 V 1 D 1 * S 2 ) * .
S 3 ( S 3 ) * = ( S 1 V 1 D 1 * S 2 ) ( S 1 V 1 D 1 * S 2 ) * = S 1 V 1 D 1 * S 2 S 2 * D 1 V 1 * S 1 * = S 1 V 1 D 1 * D 1 V 1 * S 1 * = S 1 V 1 V 1 * S 1 * = I .
( E o u t e j ϕ ) ( E o u t e j ϕ ) * = S 3 V 3 D 3 * ,
E i n ( E i n ) * = ( S 2 V 2 D 2 * ) ( S 2 V 2 D 2 * ) * = S 2 V 2 D 2 * D 2 V 2 * S 2 * = S 2 ( V 2 V 2 * ) S 2 * .
e j ϕ = [ e j ϕ 1 , e j ϕ 2 , , e j ϕ M o u t ] T [ e j ϕ 1         e j ϕ 2                 e j ϕ M o u t ] = φ .
( E o u t e j ϕ ) ( E o u t e j ϕ ) * = ( φ E o u t ) ( φ E o u t ) * = φ E o u t ( E o u t ) * φ * .
E o u t = S 4 V 4 D 4 * ,
E o u t ( E o u t ) * = ( S 4 V 4 D 4 * ) ( S 4 V 4 D 4 * ) * = S 4 V 4 D 4 * D 4 V 4 * S 4 * = S 4 ( V 4 V 4 * ) S 4 * .
( E o u t e j ϕ ) ( E o u t e j ϕ ) * = φ E o u t ( E o u t ) * φ * = φ S 4 V 4 D 4 * ( S 4 V 4 D 4 * ) * φ * = φ S 4 V 4 D 4 * D 4 V 4 * S 4 * φ * = ( φ S 4 ) ( V 4 V 4 * ) ( φ S 4 ) * .
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.