Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

High order statistics based blind deconvolution of bi-level images with unknown intensity values

Open Access Open Access

Abstract

We propose a novel linear blind deconvolution method for bi-level images. The proposed method seeks an optimal point spread function and two parameters that maximize a high order statistics based objective function. Unlike existing minimum entropy deconvolution and least squares minimization methods, the proposed method requires neither unrealistic assumption that the pixel values of a bi-level image are independently identically distributed samples of a random variable nor tuning of regularization parameters. We demonstrate the effectiveness of the proposed method in simulations and experiments.

©2010 Optical Society of America

1. Introduction

Images acquired from imaging devices are often blurred due to out-of-focus, motion of imaging devices and/or objects, etc. Since a blurred image can be modeled by the convolution of a true image and a point spread function (PSF), deconvolution using the PSF is necessary to restore the true image from the blurred acquired image. If the PSF is not known, the deconvolution becomes more challenging blind deconvolution. Since point spread functions are not completely known in many deconvolution problems, blind deconvolution has been studied intensively. Existing blind deconvolution methods include maximum likelihood type methods [1–5], linear filtering methods [6–10], etc. See excellent review articles for the details of the existing methods [11–13].

In our point of view, blind deconvolution methods may be classified into two categories. One includes linear methods that attempt to restore true image by some linear filtering of the acquired image. Such methods often iteratively determine the linear filter by maximizing an objective function based on the a prori information about the true image such as support size [6] and pixel intensity values [7–9]. The other category includes nonlinear methods that seek both an image and a PSF that maximize some objective function such as a likelihood function [1–5]. Since the maximization problem is a well known ill-posed problem in the sense that the solution is not unique (interchanging the PSF and the image would yield the same likelihood value), one is required to incorporate a priori information about the PSF and/or true image to regularize a solution. One of the most widely used regularizing methods is adding a penalty function based on the a priori information to the objective function. Since both the linear and nonlinear methods rely on the a priori information about true image and/or true PSF heavily, an effective incorporation of the information is crucial for successful blind deconvolution.

In this article, we investigate a blind deconvolution method for images that have only two intensity values (bi-level image) such as barcode, text and license plates that are found in many engineering applications [10,14,15]. There have been several investigations on the blind deconvolution of bi-level images based on the a priori information that true image has only two intensity values. For example, there have been nonlinear methods such as total variation and double-well function based regularization method [3], iterated quadratic programming method [4] and parameterized model based method [5]. In addition, linear filtering methods using three step iteration algorithm (TSIA) [8] and minimum entropy method (MED) [9] have been studied. The existing methods have their own drawbacks. For example, the total variation and double-well function based method [3] and iterated quadratic programming based method [4] assume that two intensity values are known, which is unrealistic in practice. In addition, the method in [3] requires tuning of two regularization parameters, which is difficult to automate. Another nonlinear method based on the parameterized barcode model can be applied only for 1D barcode images [5]. The existing linear methods also have drawbacks such as slow convergence [8] and noise sensitivity [8, 9].

To overcome the drawbacks of the existing methods, we study a novel linear method for the blind deconvolution of bi-level images. We focus on the linear method since it has advantages of not requiring tuning of regularization parameters. Typically, the nonlinear methods require two regularization parameters (one for regularzation function for image, the other for PSF) and the performance of the methods greatly depends on the selection of the regularization parameters. Moreover, the automatic tuning of the parameters is challenging since the selection of the regularization parameters should depend on an acquired image and unknown true PSF. In addition, to our knowledge, there have been no nonlinear blind deconvolution method for bi-level images with two unknown intensity values.

Linear deconvolution methods for bi-level images often use high order statistics as an objective function to determine an optimal filter since the second order statistics is not able to measure the binariness of an image [8,9,12]. High order statistics such as cumulants have been widely used [9,12] for the deconvolution of independently and identically distributed (i.i.d.) images (no correlation between pixel values) regardless of whether true images are binary or not. The method, also known as minimum entropy deconvolution (MED) method, has a theoretical basis that the magnitude of a normalized cumulant of a linear combination of a random variable (RV) is always smaller than that of the RV. By virtue of that, one can expect that the sample estimation of the normalized cumulant is maximized when the blur (i.e. linear combination) is removed [12]. Since the sample estimation of the fourth order normalized cumulant (often called kurtosis, which is one of the most widely used normalized cumulants for blind deconvolution) reaches upper bound when deconvolved signal has all zeros but one nonzero value, the method is also known as a method that seeks “simple or sporadic structure” signal [17]. Although the method can be used effectively for signals that have zero (or small) correlation such geophysical signals [12, 16, 17], the performance is limited for the deconvolution of an image since true image has neither zero correlation nor “simple or sporadic structure” in general.

To apply the principle of the MED method for the deconvolution of a bi-level image, a method that maximizes the sample fourth order normalized cumulant of the horizontal and vertical differences of an acquired image was investigated [9]. The method is based on the assumption that the horizontal and vertical differences in true bi-level image have more “sporadic or simple structure” than that of a blurred image. We call this method difference MED (DMED) method to discriminate it from the usual MED method. Although the method is intuitively plausible, it has several drawbacks. Firstly, it is not guaranteed that the sample kurtosis of the difference image is maximized when the deconvolved image is bi-level since the horizontal and vertical differences of a bi-level image are not i.i.d. realizations of a RV. Therefore, there might exist a filter whose output is significantly different from a bi-level image but has higher kurtosis than that of the bi-level image. In addition, the DMED method may amplify the effect of noise due to the differencing operations. To overcome the drawbacks of the DMED method, a method that minimizes a high order statistics based objective function using three step iteration approach was investigated [8]. However, the method is slow [4] and sensitive to noise [8].

To overcome the drawbacks of the existing methods, we investigate a novel method whose objective function is minimized when deconvolved image is bi-level. The method searches a Wiener-like filter (known to be more robust to noise [7]) and the mid-level of two unknown intensity values during the optimization of a high order statistics based objective function. Unlike the DMED method, the proposed method neither assume that the pixels of true image (or difference image) are i.i.d. realizations of a RV nor compute differences of an acquired image. By virtue of that, we expect that the performance of the proposed method should be better than that of the DMED method.

2. Problem formulation

We model an M×N image acquired from an imaging device by the convolution of true image and true PSF plus additive noise as follows:

y(m,n)=ΣjΣkh(nj,mk)x(j,k)+w(m,n),m=1,...M,n=1,...,N,

where y(m,n), h(m,n), x(m,n) and w(m,n) represent blurred acquired image, true PSF, true image and additive noise, respectively. The pixel value of true image at each discrete grid has one of two unknown intensity values β1 and β2 such that x(m,n) ∈ {β1, β2}. The goal of blind deconvolution is the accurate estimation of the true image x = [x(1,1), x(1,2), ...., x(M,N)] from y = [y(1,1), y(1,2), ...., y(M,N)] without knowledge of h = [h(1,1), h(1,2), ...., h(M,N)]. To achieve the goal, perhaps, the most natural approach is the following penalized nonlinear maximum likelihood type method:

(x̂,ĥ)=argminx{β1,β2},hL(x,h;y)+λ1R1(x)+λ2R2(h),

where L(·, ·) is the likelihood function, R1(x) and R2(h) are regularization functions for image x and PSF h, λ1 and λ2 are associated regularization parameters, respectively. Note that solving the problem defined in Eq. (2) is very difficult since it requires a binary optimization with two unknown binary values. In addition, automatic tuning of the two regularization parameters is also challenging. To our knowledge, no method has been proposed to solve the problem defined in Eq. (2), although there exists a method based on Eq. (2) for the case that two intensity values are known [3]. However, the method is not applicable in practice since the two intensity values are not known in general.

As an alternative, one may apply the least squares minimization (LSM) method based on the Gaussian likelihood function without binary optimization [18]. The method is based on the Gaussian likelihood and two regularization functions as follows:

(x̂,ĥ)=argminx,hΣn,my(n,m)h(n,m)*x(n,m)2+λ1R1(x)+λ2R2(h),

where * is the 2D convolution operator. One of the most widely used regularization functions is a smoothness penalty function defined as follows:

R1(x)=Σn,m(x(n+1,m)2x(n,m)+x(n1,m))2+(x(n,m+1)2x(n,m)+x(n1,m1))2
R2(h)=Σn,m(h(n+1,m)2h(n,m)+h(n1,m))2+(h(n,m+1)2h(n,m)+h(n1,m1))2

Note that the method defined in Eq. (3) does not require binary optimization. However, the method still suffers from cumbersome tuning of the two regularization parameters. In addition, regularization function such as the smoothness penalty is not based on the a priori information that true image is bi-level image. To our knowledge, effective regularization functions specialized for bi-level images with two unknown intensity values have not been investigated yet.

Unlike the nonlinear method in Eq. (2), linear methods attempt to restore the true image by determining an optimal filter that maximizes an objective function as follows:

f̂=argmaxfΦz(f;y),

where Φz(·) is some objective function that measures the degree of binariness of the filtered image z defined as follows:

z(m,n;f)=ΣjΣkf(mj,nk)y(j,k),m=1,...M,n=1,...,N,

where f = [f(1,1), f(1,2), ⋯, f(M,N)] are the coefficients of a linear filter. In other words, the method seeks a filter whose output image is the most consistent with the a priori information that the true image has only two intensity values. Often it is assumed that the PSF h(m,n) in Eq. (1) has stable inverse filter f(m,n) such that f(m,n)**h(m,n) = δ(mmo,nno), where mo and no are some shifts in the spatial domain. Clearly, the assumption is not always valid since many PSFs (such as PSF for out-of-focus blur, motion blur) do not have stable inverse. Even if the PSF is invertible, the method still suffers from amplifying the noise since the inverse filter is a high pass filter in nature. Nevertheless, several researchers have reported that the methods perform well under modest noise levels [6–9]. Note that for most cases, the linear filtering methods do not require regularization functions.

3. MED based methods

The MED method determines an optimal filter by maximizing the following objective function [17]:

f̂=argmaxf1MNΣm,n(z(m,n;f))4(1MNΣm,n(z(m,n;f))2)2.

The objective function defined in Eq. (7) can be thought as the sample estimation of the fourth order normalized cumulant (kurtosis) of a zero mean random variable (RV) Z (except for correction term -3 for making the kurtosis of a Gaussian RV zero [12]), provided that z(m,n; f), m = 1, …M, n = 1, …N are i.i.d. samples of Z. The objective function reaches the maximum value MN when the signal z(m,n; f) has all zero values but only one nonzero value. Due to that, the MED method is known as a method that seeks “simple or sporadic structure” in deconvolved signal [17]. The MED method has been effectively used for the deconvolution of geophysical signals such as seismic signals, mainly due to that ideal geophysical signals have most zero values but small numbers of spikes [16, 17].

In addition to the previous justification, it has been known that the magnitude of the normalized cumulant of a linear combination of a RV is always smaller than that of the RV [12]. Based on the fact, if the pixel intensity values of true image are i.i.d. samples of a RV, ideal inverse filter (which restores the i.i.d. sample values) would yield maximum magnitude of the normalized cumulants.

However, it is difficult to apply the method for the deconvolution of images since adjacent pixels in true image have non-negligible correlation. In addition, it is clear that seeking “simple or sporadic structure” in deconvolved image is not a good strategy in general. To avoid the difficulty, based on the observation that differences between adjacent pixels of true bi-level image might be more “sporadic (or spiky)” than that of the blurred image, the DMED method seeks a filter that maximizes the sample kurtosis of horizontal and vertical differences [9]. The DMED method is defined as follows:

f̂=argmaxfΣnΣmdx4(m,n;f)(Σmdx2(m,n;f))2+ΣmΣndy4(m,n;f)(Σndy2(m,n;f))2,

where dx(m,n; f) = z(m,n; f)−z(m−1, n; f) and dy(m,n; f) = z(m,n; f)−z(m,n−1; f).

Although the method is plausible, it has two drawbacks. Firstly, since the pixel values of the difference image are not i.i.d. realizations of a RV even for a noise free image, there might exist a filter whose output difference image is very “sporadic” but far from the difference of a bi-level image. Therefore maximizing the objective function defined in Eq. (8) might yield unrealistic image. Secondly, due to difference operation, it may amplify the effect of noises. Due to that, the method was only tested for noise free image in the previous investigation [9].

4. Proposed method

To overcome the drawbacks of the MED based methods in applying for blind deconvolution of bi-level images, we propose a novel linear filtering method based on the normalized cumulants. We pay attention to the following fact. If a deconvolved image using a linear filter is an ideal bi-level image, there exists a constant (which is the mid-level of two levels) that satisfies the following relationship:

z(n,m;f~)α~=k,(m,n),

where k is some constant. Based on the fact, we design an objective function with variables f and α that is minimized at f = and α = . To do that, we consider the following high order statistics based objective function with variables f and α:

Φ(f,α)=1MNΣm,n(z(m,n;f)α)4(1MNΣm,n(z(m,n;f)α)2)2.

The objective function in Eq. (10) reaches its lowest bound at f = and α = that satisfy the condition in Eq. (9) [12]. (The proof is a simple application of Hölder’s inequality.) Note that the objective function in Eq. (10) is minimized at f = and α = regardless of whether true bi-level image is an i.i.d. image or not. The condition in Eq. (9) can be satisfied when the deconvolved image is a bi-level image or a constant image. Since a constant image might be generated when all filter coefficients are zero, we normalize filter coefficients in a way such that the sum of the coefficients should be unity to prevent such trivial minimizer from occurring [8].

Although one may effectively restore true bi-level image by minimizing the objective function defined in Eq. (10) for the cases of invertible PSF and noise free image, the performance of such method might deteriorate since neither PSF is invertible, nor is noise power negligible in general. We attempt to improve the performance of the method by adopting Wiener-like filter that is known to be robust to noise [11, 19]. The Wiener filter is the well known linear least squares error (LLSE) filter [20] (provided that the PSF and power spectrums of the signals and noises are known) regardless of whether the PSF is invertible or not. If the power spectrums are unknown, the pseudo Wiener filter using noise-to-signal ratio (NSR) is often used [21]. We adopt the pseudo Wiener filter that generates filtered image as follows [19]:

z(m,n;h,η)=1{H*(ωx,ωy)H(ωx,ωy)2+ηY(ωx,ωy)},

where −1{·} is the 2D inverse Fourier transform operator, H(ωxy) is the Fourier transform of PSF h, and η is the NSR. Since both the PSF and NSR are unknown, we search the PSF h and NSR η that minimize the objective function defined in Eq. (10). Note that we allow negative pixel values in restored images since ultimate goal of blind deconvolution of bi-level images is an accurate estimation of the bi-level images after binarization. Negative pixel values do not affect the performance of the binarization and were allowed in a previous investigation on bi-level image restoration [4].

In summary, using the objective function defined in Eq.(10) and filtered image defined in Eq. (11), we search the PSF h, unknown mid-level α and NSR η. Further, we limit search space by incorporating additional information about the mid-level of true bi-level image and the range of η. It would be reasonable to assume that the mid-level is between the maximum and the minimum of the observed image and signal-to-noise ratio is in some range (e.g.,between 10dB and 100dB). Incorporating such information would be helpful to determine a more realistic solution. Finally, the propose method is defined as follows:

(ĥ,α̂,η̂)=argminh,α1αα2,η1ηη2Φz(h,α,η),

where α1, α2 and η1, η2 are the lower and upper bounds for α and η, respectively, and Φz(h, α, η) is defined as follows:

Φz(h,α,η)=1MNΣm,n(z(m,n;h,η)α)4(1MNΣm,n(z(m,n;h,η)α)2)2.

One may solve the constrained optimization problem defined in Eq. (12) using a constrained optimization algorithm. We solve the problem using a gradient based trust-region-reflective method [22].

Unlike the conventional MED and DMED method, the proposed method is not based on unrealistic assumption that the intensity values of pixels are i.i.d. realizations of a RV. Without the unrealistic assumption, it is guaranteed that the objective function of the proposed method is minimized when the deconvolved image is a bi-level image. In addition, we expect that the proposed method is robust to noise since not only it adopts Wiener-like filter but also it performs no noise amplifying differencing operation.

5. Results

5.1. Simulations

We conducted simulation studies to evaluate the performance of the proposed method in comparison with the DMED method. Figure 1(a) and Fig. 1(b) show true QR barcode image and text image that were used for simulation. The sizes of the QR barcode image and the text image were 290×290 and 290×468, respectively.

 figure: Fig. 1.

Fig. 1. Images for simulation: (a) true barcode image; (b) true text image.

Download Full Size | PDF

We synthesized blurred barcode image and text image by convolving the true images with a 2D separable Gaussian PSF that has the same standard deviation σ for both x and y directions. Note that the 2D Gaussian function is often used for modeling out-of-focus blur [5]. We also added white Gaussian noise to the synthesized blurred images to simulate the effect of noise. Figure 2(a) shows a synthesized noisy blurred barcode image (using PSF with σ = 6 pixels) whose SNR is 25 dB.

To quantify the degree of the similarity between an image and the true image, we defined two measures, the correlation coefficient (COR) between the image and the true image and bit error rate (BER). We computed the BER as the percentage of error pixels after the binarization of the image using a global threshold that was determined by the well-known Otsu’s method [21]. The COR value of the image shown in Fig. 2(a) was 0.786 and the BER of its binarization shown in Fig. 2(b) was 12.63%. Note that it is difficult to decode the barcode since the binarized image without deblurring has many error pixels.

To restore the true image from the noisy blurred image, we applied the DMED method and the proposed method to the image shown in Fig. 2(a). We applied 23×23 separable and symmetric filter for linear filter f for the DMED method and PSF h for the proposed method. If the support size of PSF is too small, then the blind deconvolution is not able to estimate true PSF while large support size make optimization complicated. Therefore, we determined the support size considering the maximum blur size that we attempt to compensate for. Note that there are many blind deconvolution methods based on the correct support size information. Those methods are prone to the errors in the support size of the PSF [11]. Unlike the methods, the proposed method does not require accurate knowledge of the correct support size. For example, we use the same support size for all the different blur point spread functions in the subsequent simulations and experiments. We defer investigating automatic determination of the support size to future study. We initialized the coefficients of the filter and PSF by delta function, i.e., assumed there is no blur in an acquired image. For the proposed method, we determined α1, α2 by the minimum and the maximum value of the observed image and η1, η2 by corresponding SNR 10dB and 100dB, respectively. For comparison purpose, we also applied the LSM method defined in Eq. (3) and Eq. (4). We manually tuned two regularization parameters λ1 and λ2 in a way such that the BER of the LSM method is minimized.

Figure 2(c) (COR 0.372) and Fig. 2(d) (BER 39.76%) show deblurred image by the DMED method and its binarization, respectively. (For display purpose, we normalize the intensity values in figures in this paper.) As shown in the images, the performance of the DMED method was not satisfactory. The COR value of the image shown in Fig. 2(c) was smaller than that of the image without deblurring. In addition, the BER value of the binarized image (shown in Fig. 2(d)) was larger than that of the binarization without deblurring.

The LSM method showed better performance than the DMED method. Figure 2(e) (COR 0.813) and Fig. 2(f) (BER 10.06 %) show the restored image using the LSM method and its binirization, respectively. As shown in the figures, the LSM method was able to improve the quality of the blurred image.

Compared with the DMED method and the LSM method, the proposed method showed significantly improved results. Figure 2(g) (COR 0.901) and Fig. 2(h) (BER 3.86%) show deblurred image by the proposed method and its binarization. As shown in the figures, it is clear that the proposed method significantly improved the quality of the observed image.

We believe that the poor performance of the DMED method is originated from its objective function. The DMED objective function value of the image shown in Fig. 2(c) was 7.95 while that of the image shown in Fig. 2(e) was 6.71. Apparently, more bi-level like image in Fig. 2(e) has smaller objective function value than that of the unrealistic image shown in Fig. 2(c). This is due to the fact that the DMED method seeks only more “simple and sporadic” structure in difference image regardless of binariness in deblurred image.

We also applied the three methods for the deconvolution of the text image shown in Fig. 3(a) that was generated by convolving the true text image shown in Fig. 1(b) with 2D separable Gaussian PSF (σ = 5 pixels). The SNR of the image was 35 dB and the COR value was 0.751. Figure 3(b) shows the binarization of the image in Fig. 3(a) (BER 6.69%). Figure 3(c) (COR 0.811) and Fig. 3(d) (BER 4.93%) show deblurred image and its binarization using the DMED method, respectively. Compared with the results for the barcode image, the performance of the DMED method was better as it was able to improve the quality of the observed image. We think this is due to that the edge image of the text image has more sporadic structure than that of the barcode image. For the text image, the DMED method outperformed the LSM method. As shown in the restored image by the LSM method in Fig. 3(e) (COR 0.754) and its binarization in Fig. 3(f) (BER 6.67%), the restored images by the LSM method were almost the same as blurred images.

 figure: Fig. 2.

Fig. 2. Barcode images: (a) noisy blurred barcode image (σ=6 pixels, SNR=25dB, COR=0.786); (b) binarization of Fig. 2(a) (BER=12.63%); (c) deblurred barcode image using the DMED method (COR=0.372); (d) binarization of Fig. 2(c) (BER=39.76%); (e) deblurred barcode image using the LSM method (COR=0.813); (f) binarization of Fig. 2(e) (BER=10.06%); (g) deblurred barcode image using the proposed method (COR=0.901); (h) binarization of Fig. 2(f) (BER=3.86%).

Download Full Size | PDF

Although the performance of the DMED method was better than that of the LSM method, the DMED method was no better than the proposed method. Figure 3(g) (COR 0.894) and Fig. 3(h) (BER 2.18%) show deblurred image and its binarization using the proposed method. The COR value of the deblurred image using the proposed method was significantly larger than that of using the DMED method and the BER value of the binarized image using the proposed method is less than half of that using the DMED method.

The previous results show the performances of the DMED method, the LSM method and the proposed method for just one noise realization and a fixed amount of blur. To evaluate the performance of each method more thoroughly, we repeated the deblurring procedure for 50 different noise realizations for different noise powers and different amounts of blur (e.g., different σ values in PSF). Table 1 and Table 2 show average COR and BER values of the 50 noise realizations for the deconvolution of the barcode image, respectively. As shown in the tables, the performance of the proposed method was significantly better than that of the DMED method for every SNR and σ values. As a matter of fact, deblurred images by the DMED method were worse than observed images in the sense that the COR and BER values after applying the DMED method were worse than that of blurred images. On the contrary, the proposed method was able to improve the COR and BER values for every cases. For the cases of small blur and high SNR, the LSM method with manually tuned optimal parameters showed better performance than the proposed method. However, for most cases, the proposed method outperformed the LSM method. Note that we manually searched two regularization parameters (λ1 = 10−2,λ2 = 10−4) for the LSM method. It is challenging to tune the regularization parameters automatically since the optimal parameters vary depending on unknown true image, unknown PSF and unknown SNR. By comparison, the proposed method showed better performance for most cases without requiring manual tuning of any parameter.

Tables Icon

Table 1. COR values of blurred images and deblurred images by the three methods for the barcode image with different amounts of blur and SNR values (unit for SNR is dB.)

Table 3 and Table 4 show average COR and BER values for the deconvolution of the text image, respectively. Unlike the results for the barcode image, the DMED method was able to improve the quality of blurred text images. However, the performance of the proposed method was better than that of the DMED method for every cases except for noisy and severely blurred images (e.g., SNR=20dB, 15dB and σ = 4, σ = 5 pixels) where the COR and BER values of the DMED method were slightly better than that of the proposed method. We do not think the results of such cases are meaningful for performance comparison since characters in restored images using both methods were not recognizable (e.g. BER is higher than 5%). The performance of the LSM methods was slightly better than that of the proposed method for the cases of the smallest blur (e.g., σ = 2 pixels). (Note that we had to search optimal parameters (λ1 = 10−1,λ2 = 105) for the text image again since that were not same as the optimal parameters for the barcode image.) However, for most cases, the proposed method outperformed the DMED method and the LSM method as shown in the tables.

 figure: Fig. 3.

Fig. 3. Text images: (a) noisy blurred text image (σ=5 pixels, SNR=35dB, COR=0.751); (b) binarization of Fig. 3(a) (BER=6.69%); (c) deblurred text image using the DMED method (COR=0.811); (d) binarization of Fig. 3(c) (BER=4.93%); (e) deblurred text image using the LSM method (COR=0.754); (f) binarization of Fig. 3(e) (BER=6.67%); (g) deblurred text image using the proposed method (COR=0.894); (h) binarization of Fig. 3(g) (BER=2.18%).

Download Full Size | PDF

Tables Icon

Table 2. BER values of blurred images and deblurred images by the three methods for the barcode image with different amounts of blur and SNR values (unit for BER is %.)

Tables Icon

Table 3. COR values of blurred images and deblurred images by the three methods for the text image with different amounts of blur and SNR values (unit for SNR is dB.)

Tables Icon

Table 4. BER values of blurred images and deblurred images by the three methods for the text image with different amounts of blur and SNR values (unit for BER is %.)

One may argue that the performance of the LSM method can be improved by other regularization functions (possibly by a novel regularization function that is specialized for bi-level images or total variation [23]). We do not intend to claim that the proposed method outperforms every LSM methods. The purpose of the comparison is to show the effectiveness of the proposed method that does not require tuning of regularization parameters. More thorough comparison with possible nonlinear methods is deferred to future study.

We also tested the performance of the proposed method for non-Gaussian blur. We conducted simulation studies using a non-minimum phase non-causal auto-regressive (AR) filter blur that was used to test the performance of blind deconvolution of bi-level images in a previous investigation [8]. The Fourier transform of a noisy blurred image by the AR filter blur is defined as follows [8]:

Y(ωx,ωy)=(1ρ4)(1ρejωx)(1ρejωx)(1ρejωy)(1ρejωy)X(ωx,ωy)+W(ωx,ωy),

where ρ represents correlation between adjacent pixels and W(ωxy) does the Fourier transform of Gaussian noise. Figure 4(a) and Fig. 4(b) show blurred barcode and text image (SNR=30dB, ρ = 0.82 for barcode and ρ = 0.80 for text) using the AR filter. Figure 4(c) and Fig. 4(d) show the binarization of the blurred barcode and text image, respectively. We applied the DMED, the LSM and the proposed method for the blind deconvolution of the blurred images. Figure 4(e) and Fig. 4(f) show binarized images after applying the DMED method while Fig. 4(g) and Fig. 5(h) do those after applying the LSM method. (For brevity, we do not show deconvolved images.) As shown in the figures, the performance of the DMED method for the barcode image and the performance of the LSM method for the text image were very poor as the BER values after applying the methods were higher than those without deconvolution. Although the DMED and LSM method were able to improve the text image and the barcode image, respectively, the proposed method outperformed the methods as shown in the banarized barcode image (Fig. 4(i)(BER 1.55%)) and binarized text image (Fig. 4(j)(BER 1.96%)).

To evaluate the statistical properties of the three methods for the blurred images by the AR filter, we repeated simulations for 50 times for different amounts of blur (i.e. different ρ values) and different SNR values. Table 5 and Table 6 show average BER values of the three methods. (For brevity, we do not show COR values.) As shown in the tables, the proposed method outperformed the DMED method and the LSM method.

Tables Icon

Table 5. BER values of blurred images and deblurred images by the three methods for the barcode image blurred by AR filter with different amounts of blur and SNR values (unit for BER is %.)

Tables Icon

Table 6. BER values of blurred images and deblurred images by the three methods for the text image blurred by AR filter with different amounts of blur and SNR values (unit for BER is %.)

Note that the performance of non-convex optimization based methods (such as the DMED method and the proposed method) depends on initial estimate of parameters since optimization may converge to the nearest local minimum from the initial guess [24]. It is very difficult to analyze the robustness to initial estimate since the effect of initial estimate can be different for different images and different noise realizations. During our simulations, we initialized PSF by delta function (i.e. no blur) for images blurred by Gaussian function and barcode images blurred by AR filter, and optimization converged to global minimum. In addition, we tested different initial PSFs (Gaussian PSFs with different σ values) and found optimization converged to global minimum for certain ranges of σ values. Note that the ranges are different for different amounts of blur and different SNR values. For blurred text images by the AR filter, we repeated optimization using initial Gaussian PSFs with different σ values and determined image that has the smallest objective function value. As a matter of fact, the strategy is one of the most widely used simple global optimization strategies [25]. We do not think that the existence of local minima in the proposed objective function makes the proposed method less effective since we were able to find global minimum by the simple strategy. One may also apply global optimization algorithms such as simulated annealing method and genetic algorithm [24]. We defer more thorough investigation on applying the global optimization methods to future study.

 figure: Fig. 4.

Fig. 4. Images for AR blur: (a) barcode; (b) text; (c) binarization of Fig. 4(a)(BER 8.48%); (d) binarization of Fig. 4(b)(BER 6.14%); (e) barcode by DMED (BER 41.9%); (f) text by DMED(BER 2.92%); (g) barcode by LSM(BER 6.21%); (h) text by LSM (BER 4.66%); (i) barcode by proposed (BER 1.55%) (j) text by proposed (BER 1.96%).

Download Full Size | PDF

5.2. Real image

We also tested the performance of the proposed method in comparison with the DMED method for real images. We acquired real barcode and text images under normal illumination using a digital camera (Canon EOS-40D) without auto focusing. Figure 5(a) and Fig. 5(b) show the real barcode image and its binarization, respectively. As shown in the figures, due to blur and noise present in the acquired image, the binarized image has many error pixels. Figure 5(c) and Fig. 5(d) show restored image and its binirization using the DMED method. The performance of the DMED method was not satisfactory as the results of the restoration were even worse than the acquired images without deblurring. Figure 5(e) and Fig. 5(f) show restored image and its binarization using the LSM method. For the LSM method, we searched best parameters based on the subjective evaluation of the quality of restored image. Unlike the DMED method, the LSM method was able to improve the quality of the acquired image. Figure 5(e) and Fig. 5(f) show deblurred image using the proposed method and its binarization, respectively. Based on subjective evaluation, we believe that the proposed method showed better performance than the DMED and the LSM method. Although it is not possible to quantify the degree of the improvement since ground truth image is not available, it is clear that the binarized image using the proposed method (Fig. 5(h)) is more similar to the image shown in Fig. 1(b) than other binzarized images such as Fig. 5(d) and Fig. 5(f).

We also tested the performance of the three methods using a real text image. Figure 6(a) and Fig. 6(b) show the real text image and its binarization, respectively. Figure 6(c) and Fig. 6(d) show deblurred image and its binarization using the DMED method while Fig. 6(e) and Fig. 6(f) show those using the LSM method. Unlike the real barcode image case, the DMED method was able to improve the acquired image as shown in Fig. 6(c) and Fig. 6(d). As pointed out before, this phenomenon is originated from the fact that the edge image of the true text image has more “simple and sporadic” structure than the barcode image. Figure 6(g) and Fig. 6(h) shows restored image and its binarization using the proposed method. Based on the subjective evaluation of Fig. 6(g) and Fig. 6(h) in comparison with Fig. 6(c), Fig 6(d), Fig. 6(e) and Fig. 6(f), we believe that the proposed method outperformed the DMED method and the LSM method.

6. Conclusions

We have proposed a novel linear blind deconvolution method for bi-level images. The proposed method searches mid-level of two intensity values, PSF and noise-to-signal ratio by minimizing a high order statistics based objective function that reaches its lowest bound when deconvolved image is a bi-level image. Unlike conventional methods, the proposed method requires neither unrealistic assumptions such as image pixel values are i.i.d. samples of a random variable nor tuning of regularization parameters. We demonstrated the effectiveness of the proposed method in simulations and experiments using barcode and text images.

 figure: Fig. 5.

Fig. 5. Real barcode image and deblurred images: (a) real noisy blurred barcode image; (b) binarization of image shown in Fig. 5(a); (c) deblurred real barcode image using the DMED method; (d) binarization of image shown in Fig. 5(c); (e) deblurred real barcode image using the LSM method; (f) binarization of image shown in Fig. 5(e); (e) deblurred real barcode image using the proposed method; (f) binarization of image shown in Fig. 5(g).

Download Full Size | PDF

 figure: Fig. 6.

Fig. 6. Real text image and deblurred images: (a) real noisy blurred text image; (b) binarization of image shown in Fig. 6(a); (c) deblurred real text image using the DMED method; (d) binarization of image shown in Fig. 6(c); (e) deblurred real text image using the LSM method; (f) binarization of image shown in Fig. 6(e); (g) deblurred real text image using the proposed method; (h) binarization of image shown in Fig. 6(g).

Download Full Size | PDF

Acknowledgments

The authors are very grateful to the anonymous reviewers for thorough review of the paper and many insightful suggestions. The authors are also grateful to Sohyun Ahn for helping us prepare this paper. This work was supported by the Korean Research Foundation Grant funded by the Korean Government (KRF-2008-331-D00419) and Korean Science Engineering Foundation (KOSEF R17-2008-041-01001-0) funded by the Korean Government.

References and links

1. T. J. Holmes, “Blind deconvolution of quantum-limited incorehent imagery:maximum-likelihood approach,” J. Opt. Soc. Am. A 9, 1052–1061 (1992). [CrossRef]   [PubMed]  

2. D. A. Fish, A. M. Brinicombe, E. R. Pike, and G. Walker, “Blind deconvolution by means of the Richardson-Lucy algorithm,” J. Opt. Soc. Am. A 12, 58–65 (1995). [CrossRef]  

3. S. Esedoglu, “Blind deconvolution of bar code signals,” Inverse Probl. 20, 121–135 (2004). [CrossRef]  

4. E. Y. Lam, “Blind bi-level image restoration with iterated quadratic programming,” IEEE Trans. Circ. Syst. Part 2 52, 52–56 (2007). [CrossRef]  

5. J. Kim and H. Lee, “Joint nonuniform illumination estimation and deblurring for bar code signals,” Opt. Express 17, 14817–14837 (2007). [CrossRef]  

6. D. Kundur and D. Hatzinakos, “A novel blind deconvolution scheme for image restoration using recursive filtering,” IEEE Trans. Signal Process. 45, 375–390 (1998). [CrossRef]  

7. G. R. Ayers and J. C. Dainty, ‘Iterative blind deconvolution method and its application,” Opt. Lett. 13, 547–549 (1998). [CrossRef]  

8. T. Li and K. Lii, “A joint estimation approach for two-tone image deblurring by blind deconvolution,” IEEE Trans. Image Process. 11, 847–858 (2002). [CrossRef]  

9. H. Wu, “Minimum entropy deconvolution for restoration of blurred two-tone images,” Electronics Letters 26, 1183–1184 (1990). [CrossRef]  

10. N. Miura, N. Baba, S. Isobe, M. Noguchi, and Y. Norimoto, “Binary star reconstruction with use of the blind deconvolution method,” J. Mod. Opt. 39, 1137–1146 (1992). [CrossRef]  

11. D. Kundur and D. Hatzinakos, “Blind image deconvolution,” IEEE Trans. Image Process. 2, 223–235 (1993).

12. J. A. Cadzow, “Blind deconvolution via cumulant extrema,” IEEE Signal Processing Magazine., 24–41 (1996). [CrossRef]  

13. P. Campisi and K. Egiazarian Eds., Blind image deconvolution: Theory and applications, (CRC, New York, 2007). [CrossRef]  

14. H. Lee and J. Kim, “Retrospective correction of nonuniform illumination on bi-level images,” Opt. Express 15, 23880–23893 (2009). [CrossRef]  

15. Y. Shen, E. Y. Lam, and N. Wong, “Binary image restoration by positive semidefinite programming,” Opt. Lett. 32, 121–123 (2007). [CrossRef]  

16. M. D. Sacchi, D. R. Velis, and A. H. Comingues, “Minimum entropy deconvolution with frequency-domain constraints,” Geophysics 59, 938–945 (1994). [CrossRef]  

17. D. Donoho, “On minimum entropy deconvolution,” Applied Time Series Analysis II, D. F. Findley ed., (Academic, New York, 1991).

18. N. F. Law and R. G. Lane, “Blind deconvolution using least squares minimisation,” Opt. Commun. 128, 341–352 (1996). [CrossRef]  

19. J. Kim, “Restoration of bi-level images via iterative semi-blind Wiener filtering,” Trans. KIEE 57, 1290–1294 (2008).

20. H. L. Van Trees, Detection, estimation, and modulation theory, Part 1 (Wiley, 1968).

21. R. C. Gonzalez, R. E. Woods, and S. L. Eddins, Digital image processing using MATLAB, (Prentice Hall, New York, 2002).

22. T. Mathworks, Optimization toolbox user’s guide (Mathworks Inc., 2003).

23. T. Chan and C. K. Wong, “Total variation blind deconvolution,” IEEE Trans. Image Process. 7, 370–375 (1998). [CrossRef]  

24. E. K. P. Chong and S. H. Żak, An introduction to optimization, 3rd ed. (Wiley-Interscience, New Jersey, 2008).

25. W. H. Press, S. A. Teukolsky, W. T. Vetterling, and B. P. Flannery, Numerical recipes in C++, 2nd ed. (Cambridge, 2005).

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (6)

Fig. 1.
Fig. 1. Images for simulation: (a) true barcode image; (b) true text image.
Fig. 2.
Fig. 2. Barcode images: (a) noisy blurred barcode image (σ=6 pixels, SNR=25dB, COR=0.786); (b) binarization of Fig. 2(a) (BER=12.63%); (c) deblurred barcode image using the DMED method (COR=0.372); (d) binarization of Fig. 2(c) (BER=39.76%); (e) deblurred barcode image using the LSM method (COR=0.813); (f) binarization of Fig. 2(e) (BER=10.06%); (g) deblurred barcode image using the proposed method (COR=0.901); (h) binarization of Fig. 2(f) (BER=3.86%).
Fig. 3.
Fig. 3. Text images: (a) noisy blurred text image (σ=5 pixels, SNR=35dB, COR=0.751); (b) binarization of Fig. 3(a) (BER=6.69%); (c) deblurred text image using the DMED method (COR=0.811); (d) binarization of Fig. 3(c) (BER=4.93%); (e) deblurred text image using the LSM method (COR=0.754); (f) binarization of Fig. 3(e) (BER=6.67%); (g) deblurred text image using the proposed method (COR=0.894); (h) binarization of Fig. 3(g) (BER=2.18%).
Fig. 4.
Fig. 4. Images for AR blur: (a) barcode; (b) text; (c) binarization of Fig. 4(a)(BER 8.48%); (d) binarization of Fig. 4(b)(BER 6.14%); (e) barcode by DMED (BER 41.9%); (f) text by DMED(BER 2.92%); (g) barcode by LSM(BER 6.21%); (h) text by LSM (BER 4.66%); (i) barcode by proposed (BER 1.55%) (j) text by proposed (BER 1.96%).
Fig. 5.
Fig. 5. Real barcode image and deblurred images: (a) real noisy blurred barcode image; (b) binarization of image shown in Fig. 5(a); (c) deblurred real barcode image using the DMED method; (d) binarization of image shown in Fig. 5(c); (e) deblurred real barcode image using the LSM method; (f) binarization of image shown in Fig. 5(e); (e) deblurred real barcode image using the proposed method; (f) binarization of image shown in Fig. 5(g).
Fig. 6.
Fig. 6. Real text image and deblurred images: (a) real noisy blurred text image; (b) binarization of image shown in Fig. 6(a); (c) deblurred real text image using the DMED method; (d) binarization of image shown in Fig. 6(c); (e) deblurred real text image using the LSM method; (f) binarization of image shown in Fig. 6(e); (g) deblurred real text image using the proposed method; (h) binarization of image shown in Fig. 6(g).

Tables (6)

Tables Icon

Table 1. COR values of blurred images and deblurred images by the three methods for the barcode image with different amounts of blur and SNR values (unit for SNR is dB.)

Tables Icon

Table 2. BER values of blurred images and deblurred images by the three methods for the barcode image with different amounts of blur and SNR values (unit for BER is %.)

Tables Icon

Table 3. COR values of blurred images and deblurred images by the three methods for the text image with different amounts of blur and SNR values (unit for SNR is dB.)

Tables Icon

Table 4. BER values of blurred images and deblurred images by the three methods for the text image with different amounts of blur and SNR values (unit for BER is %.)

Tables Icon

Table 5. BER values of blurred images and deblurred images by the three methods for the barcode image blurred by AR filter with different amounts of blur and SNR values (unit for BER is %.)

Tables Icon

Table 6. BER values of blurred images and deblurred images by the three methods for the text image blurred by AR filter with different amounts of blur and SNR values (unit for BER is %.)

Equations (15)

Equations on this page are rendered with MathJax. Learn more.

y ( m , n ) = Σ j Σ k h ( n j , m k ) x ( j , k ) + w ( m , n ) , m = 1 , . . . M , n = 1 , . . . , N ,
( x ̂ , h ̂ ) = argmin x { β 1 , β 2 } , h L ( x , h ; y ) + λ 1 R 1 ( x ) + λ 2 R 2 ( h ) ,
( x ̂ , h ̂ ) = argmin x , h Σ n , m y ( n , m ) h ( n , m ) * x ( n , m ) 2 + λ 1 R 1 ( x ) + λ 2 R 2 ( h ) ,
R 1 ( x ) = Σ n , m ( x ( n + 1 , m ) 2 x ( n , m ) + x ( n 1 , m ) ) 2 + ( x ( n , m + 1 ) 2 x ( n , m ) + x ( n 1 , m 1 ) ) 2
R 2 ( h ) = Σ n , m ( h ( n + 1 , m ) 2 h ( n , m ) + h ( n 1 , m ) ) 2 + ( h ( n , m + 1 ) 2 h ( n , m ) + h ( n 1 , m 1 ) ) 2
f ̂ = argmax f Φ z ( f ; y ) ,
z ( m , n ; f ) = Σ j Σ k f ( m j , n k ) y ( j , k ) , m = 1 , . . . M , n = 1 , . . . , N ,
f ̂ = argmax f 1 MN Σ m , n ( z ( m , n ; f ) ) 4 ( 1 MN Σ m , n ( z ( m , n ; f ) ) 2 ) 2 .
f ̂ = argmax f Σ n Σ m d x 4 ( m , n ; f ) ( Σ m d x 2 ( m , n ; f ) ) 2 + Σ m Σ n d y 4 ( m , n ; f ) ( Σ n d y 2 ( m , n ; f ) ) 2 ,
z ( n , m ; f ~ ) α ~ = k , ( m , n ) ,
Φ ( f , α ) = 1 MN Σ m , n ( z ( m , n ; f ) α ) 4 ( 1 MN Σ m , n ( z ( m , n ; f ) α ) 2 ) 2 .
z ( m , n ; h , η ) = 1 { H * ( ω x , ω y ) H ( ω x , ω y ) 2 + η Y ( ω x , ω y ) } ,
( h ̂ , α ̂ , η ̂ ) = argmin h , α 1 α α 2 , η 1 η η 2 Φ z ( h , α , η ) ,
Φ z ( h , α , η ) = 1 MN Σ m , n ( z ( m , n ; h , η ) α ) 4 ( 1 MN Σ m , n ( z ( m , n ; h , η ) α ) 2 ) 2 .
Y ( ω x , ω y ) = ( 1 ρ 4 ) ( 1 ρ e j ω x ) ( 1 ρ e j ω x ) ( 1 ρ e j ω y ) ( 1 ρ e j ω y ) X ( ω x , ω y ) + W ( ω x , ω y ) ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.