Research Article  Open Access
The Nonlocal Sparse Reconstruction Algorithm by Similarity Measurement with Shearlet Feature Vector
Abstract
Due to the limited accuracy of conventional methods with image restoration, the paper supplied a nonlocal sparsity reconstruction algorithm with similarity measurement. To improve the performance of restoration results, we proposed two schemes to dictionary learning and sparse coding, respectively. In the part of the dictionary learning, we measured the similarity between patches from degraded image by constructing the Shearlet feature vector. Besides, we classified the patches into different classes with similarity and trained the cluster dictionary for each class, by cascading which we could gain the universal dictionary. In the part of sparse coding, we proposed a novel optimal objective function with the coding residual item, which can suppress the residual between the estimate coding and true sparse coding. Additionally, we show the derivation of selfadaptive regularization parameter in optimization under the Bayesian framework, which can make the performance better. It can be indicated from the experimental results that by taking full advantage of similar local geometric structure feature existing in the nonlocal patches and the coding residual suppression, the proposed method shows advantage both on visual perception and PSNR compared to the conventional methods.
1. Introduction
The reasonable representation is the basis of many tasks in image processing. The meaning of “reasonable” is to express the important information in image with less coefficients, which is called sparse representation [1]. We always want to express the signal in a costeffective way, which could reduce the cost in signal processing. However, the emerging of sparse representation satisfied the requirements exactly. Wavelet is the landmark work, which is the most optimal expression of onedimension signal. But it shows weak ability to express the highdimension signal for its limited directions. For solving the problem, Donoho proposed the multiscale geometric analysis theory, which contains Ridgelet, Curvelet [2], Bandlet, and Contourlet [3] mainly, and then applied them to image restoration, which achieved good effect.
As for the sparse representation in image restoration, there are two main approaches, including multiscale geometric analysis (MGA) and dictionary learning. With the MGA, Donoho proposed the pioneer wavelet threshold shrinking method. After that, Yin designed a nongaussian bivariate Shrinkage function [4] under the MAP norm, which got the better performance. In 2006, as the limited directions in wavelet, Arthur proposed the NSCT transformation [5, 6] and applied it in the restoration area by Bayesian framework. But these conventional restoration models may not be accurate enough when the image is degraded seriously for their fixed overcompleted dictionary, which is not sufficient to present the abundant structures in nature images. So, a new concept is proposed, which is called dictionary learning [7–9]. Besides, many methods about how to sparse coding are proposed, such as MP [10], OMP [11], and GP [12], which is called sparse coding. Further, a method called KSVD [13, 14] is proposed by Aharon, which updates the dictionary by making SVD decomposition to the residual signal, and is applied to precision image restoration.
The goal of this paper is to research the sparse construction algorithm for image restoration based on the dictionary learning framework. We propose a nonlocal block sparse reconstruction scheme by shearlet feature vector, by which we measure the similarity between image patches and classify them for learning cluster dictionary. Then, we propose a new objective function for sparse coding to realize the high accuracy image restoration.
2. Image Restoration with Dictionary Learning
Suppose that the model among the observation signal , original signal , and noise satisfied
According to the sparse representation theory in [1], the sparse coding is equal to solving the optimization problem in (2):
In which, is the sparse dictionary of and is the coding vector. Then, we can get the original signal by . is a constant with the noise standard deviation, .
For the general restoration algorithm based on the sparse theory, dictionary is fixed. So a new design for dictionary learning called KSVD is proposed in reference [13] and good results have been achieved in image restoration. In the KSVD, we implement the sparse coding and dictionary learning simultaneously. And, the optimization task in (2) can be changed to
Elad et al. divided the image into many patches for not increasing the dictionary dimension and got the blockbased sparse coding with
In which, presents the patch whose central is . The size of patch is 7 7. They are gained by sliding a squared windows in the image. Then, we can get the restored image by averaging these patches.
Generally, KSVD learns the dictionary with some random training examples from the image, so the learned atoms show some weakness in presenting certain local structure in image, which makes the coding not sparsity enough. So, the paper has two contributions. The first is that we proposed a dictionary learning scheme based on clustering, which trained subdictionary for each cluster and then produced the universal dictionary with them. Secondly, we proposed a novel objective function for sparse coding, which added a coding residual item compared to the traditional ones. The detail for the two contributions will be introduced as follows.
3. A Novel Scheme for Image Restoration
3.1. Nonsubsampling Shearlet Transform
Nonsubsampling Shearlet [15] is a multiscale geometric analytical method advanced by Lim and Kutyniok. It consisted of laplacian pyramid and direction filter bands with multidirections and shift invariant. For its parabola scale principle, Nonsampling Shearlet has excellent performances on capturing geometrical feature and singularity in highdimension, which makes it widely used in image restoration.
3.2. Similarity Measurement
As to the many advantages with Shearlet, we propose a similarity measurement method. Implement the Shearlet transform with levels and construct the feature vector for each pixel as
Here, is the number of subband in th level and is the Shearlet coefficient. With the anisotropy feature, the coefficient for signal is large while the local geometric structure is similar to the basis function. Conversely, the coefficient is small, while the noise has the isotropy feature, so the coefficients are uniform in each primary function. Due to the above two reasons, the vector in (5) has the better antinoise ability, according to which, we take (5) to measure the similarity between the two patches. Suppose the size of patch is 5 5, we get the vector in (5) for all the 25 pixels. Then, we construct some subvector generated by the th dimension of :
In which, the presents the th element of ( is the order number of the pixel in each patch, ). Suppose and are from two patches to be measured, we compute the index as follows:
In which, is called shearlet wave coefficient ( is the coefficient standard deviation with th subband). With the formulation in (6), we can measure the similarity between two patches better than the Euclidean distance of gray in spatial domain.
3.3. Nonlocal Dictionary Learning
3.3.1. Nonlocal Dictionary Learning
The restoration algorithm consisted of two main parts that are dictionary learning and sparse coding, respectively. In this paper we improve the algorithm with the method for similarity measurement in Section 3.2. It is different to the conventional methods that we did not simply select the training patches from different example images to train the dictionary, while we select the nonlocal similar patches from the degraded image we want to restoration. And then, we cluster these patches to train the cluster subdictionary, which is more sufficient to make the universal dictionary better adapt to local geometrical feature than global KSVD. Besides, the training data in the same cluster is similar to each other, so there is no need to produce an overcompleted dictionary. So each subdictionary size is half of the size of corresponding cluster. Now we present the concrete description for learning scheme as follows.
Step 1. Implement the Shearlet transform to the image .
Step 2. Construct the Shearlet feature vector according to the formula (5) (The patch size is ).
Step 3. Calculate the standard deviation of th subband and gain the wave coefficient , and is the shearlet coefficient and the mean value of coefficients; is the total number of coefficients in th subband.
Step 4. Cluster the patches by method in [16] with the index in formula (6) and then produce the cluster ( is the cluster indicator).
Step 5. Take the patches in as the training data and learning the cluster subdictionary with KSVD.
Step 6. By cascading all the cluster dictionaries, we can get the universal dictionary for the whole image.
With the ability of capturing the local geometrical feature, the patches in the same set are highly similar by their geometry structure, which makes atoms in subdictionary have the strong adaptability to local structure. So, they can sufficiently present the local geometrical feature. Cascaded by all the subdictionaries, the universal dictionary realizes the goal that presents all kinds of features in the whole image.
Now we show a simple example to show the advantage of clusteringbased dictionary learning. We list a set of training examples:
Here, “randn” means produce a Gaussian stochastic variable, and then, we take the above two groups of training data to learn a dictionary with two atoms. We adopt the proposed algorithm and KSVD, respectively. Set the Maximum iteration number with KSVD to be 10 and we show the learned dictionary in Table 1.

As can be seen in Table 1, the atoms by proposed algorithm have the more similar structure with the training data.
3.3.2. Sparse Coding
For improving the coding accuracy, we proposed a novel optimization called coding residual suppression optimization, which is more sufficient than norm sparse coding. When dictionary is given, the norm sparse coding for original signal can be realized as follows:
In the restoration problem, we can only get the observation . So the sparse coding can be rewritten as
Comparing (9) and (10), we can see that if we want to reconstruct with high accuracy, the coding is expected to be as close as possible to the true sparse coding . So we introduce the residual coding as follows:
Then we change the object function (10) into the residual suppression form: where is the indicator of th patch. In addition, we cannot get the true sparse coding practically, so the is a good estimation of the true sparse coding and can be calculated from weighted average of sparse coding of patches in the same cluster; and are the regularization parameters. If th patch is from , we calculate :
The second term in optimization equation (12) is used to ensure the local sparsity that only part of atoms are selected for the dictionary. But, in our scheme, we select one subdictionary for each patch in a certain cluster, which means that the coding of another subdictionary is zero. So our scheme guarantees the sparsity naturally; that is, we can move the second term in (12) and rewrite it as follows:
With (14), we can compact the optimization only with coding residual constraint, which means that the coding from the observed signal by (14) is close to that from the original signal by (9).
3.3.3. Scheme Summary
The whole scheme consisted of twolevel iterative algorithm; now we present the brief steps as follows.
Initial: set the initial image ( is the degraded image); is the DCT complete dictionary and calculate the initial sparse coding for each patch with any pursuit algorithm.
Outer Loop:(a)on the (OL)th iteration: learn the dictionary by with the algorithm proposed in Section 3.3.1 ( is the maximum iterative number, );(b)compute the good estimation set for all the clusters under the .(c)Inner Loop:(i)for each patch, we get its coding by (14), which can be solved with the method in [17];(ii)repeat (i) until all the patches are processed;(d)estimate all the restoration patches by ;(e), repeat.
On one hand, the is more and more approach to the true sparse coding, which makes the residual suppression more sufficient. On the other hand, by alternating the sparse coding and dictionary learning, coding and dictionary are all improved and promote each other.
3.3.4. Parameter Selection
In conventional models, the regularization parameter is generally a constant. Hence, for making (14) more sufficient, we give the derivation of selfadaptive regularization parameter under the Bayes framework.
For written convenience, we take as the residual coding vector. Under the Bayes framework, the MAP estimation of can be computed as And the is where is the standard deviation of noise.
As for , with some statistics experiments, we gain the experience model with the i.i.d Laplacian distribution:
In which, is the th element of the residual vector of the th patch and is the standard deviation of .
Combining (15), we can obtain the following:
For a given , the optimal sparse coding can be obtained as follows:
Compared to the regularization in (14), we can set the to be the following selfadaptive form:
4. Experimental Results and Analysis
To verify the performance of the algorithm proposed in the paper, we show some contrast restoration experiments for image denoising. The contrast algorithms are, respectively, NSCT method in [6], SSBM3D in [18], MKSVD in [14], and the proposed algorithm in this paper. The noisy image is generated by adding the Gaussian noise with different standard deviation (). The size of test images is all 256 256 and we set the parameters as follows.
, (mentioned in Section 3.2); (mentioned in Section 3.3.1); , (mentioned in Section 3.3.2).
To show the objective evaluation for restoration image, we take the PSNR as the indicator for different algorithms. Meanwhile, we show part of the experiment results to view the denoising performance. Suppose and are the original image and restoration image, respectively, the definition of PSNR is
Figures 1 and 2 are the restoration performance with the standard deviation , while Figures 3 and 4 are the performance with . As for the other images we only show the PSNR value reported in Tables 2 and 3.


(a) Noisy
(b) Original
(c) NSCT
(d) SSBM3D
(e) MKSVD
(f) Proposed
(a) Noisy
(b) Original
(c) NSCT
(d) SSBM3D
(e) MKSVD
(f) Proposed
(a) Noisy
(b) Original
(c) NSCT
(d) SSBM3D
(e) MKSVD
(f) Proposed
(a) Noisy
(b) Original
(c) NSCT
(d) SSBM3D
(e) MKSVD
(f) Proposed
For each group, by the local detail image, we can get some conclusion. Though all the four methods can achieve the denoising task, there are some differences between each other. For the NSCT method, the ability of capturing structure is not adaptive to different image due to its fixed dictionary. Additionally, some scratches appeared in the restoration results, which did not exist in the original image and the PSNR is lowest among the four methods. Compared to the other method, BM3D shows advantage on PSNR value. But its restoration results are smoothed excessively, which leads to lose much detail texture. Owing to the dictionary that is learned by the random example patches from different images, the KSVD algorithm generates some visual artifacts and cannot recover the local geometry feature sufficiently. So the restoration results of KSVD are only better than the NSCT in the four methods. The proposed method in this paper shows advantage both on the restoration performance and PSNR value. Though its PSNR is lower than the BM3D, the detail recovered ability is stronger than the other three methods.
5. Conclusion
Owing to that the abundant geometry information and the selfsimilarity are the important influencers to be utilized in the image restoration by sparse representation, the paper proposed a novel scheme that restores the image by the nonlocal sparse reconstruction with similarity measurement. We cluster the patches from the degraded image by similarity measurement with Shearlet feature vector, which is good at capturing the local geometry structure in image and then take them to train the cluster dictionary and get good estimation of true sparse coding for original image. Additionally, we also show the derivation of regularization parameter under the Bayes framework. By the cluster dictionary learning and the coding residual suppression, the proposed scheme shows advantages both on the visual performance and PSNR value compared to the leading denoising methods. Besides the image denoising, the proposed model can be extensive to other restoration tasks such as deblurring and superresolution.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This work is supported by the Nation Nature Science Foundation of China nos. 61201237 and 61301095, Nature Science Foundation of Heilongjiang Province of China no. QC2012C069, and the Fundamental Research Funds for the Central Universities no. HEUCFZ1129. Meantime, this paper is funded by the Defense preresearch foundation Project of shipbuilding industry science no. 10J3.1.6.
References
 M. Alferdo, D. L. Donoho, and M. Elad, “From sparse solutions of systems of equations to sparse modeling of signals and images,” Society for Industrial and Applied Mathematics, vol. 51, no. 1, pp. 34–81, 2009. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 G. G. Bhutada, R. S. Anand, and S. C. Saxena, “Edge preserved image enhancement using adaptive fusion of images denoised by wavelet and curvelet transform,” Digital Signal Processing, vol. 21, no. 1, pp. 118–130, 2011. View at: Publisher Site  Google Scholar
 L. Shang, P.G. Su, and T. Liu, “Denoising MMW image using the combination method of contourlet and KSC shrinkage,” Neurocomputing, vol. 83, no. 15, pp. 229–233, 2012. View at: Publisher Site  Google Scholar
 S. Yin, L. Cao, Y. Ling, and G. Jin, “Image denoising with anisotropic bivariate shrinkage,” Signal Processing, vol. 91, no. 8, pp. 2078–2090, 2011. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 A. L. da Cunha, J. Zhou, and M. N. Do, “The nonsubsampled contourlet transform: theory, design, and applications,” IEEE Transactions on Image Processing, vol. 15, no. 10, pp. 3089–3101, 2006. View at: Publisher Site  Google Scholar
 J. Jia and L. Chen, “Using normal inverse Gaussian model for image denoising in NSCT domain,” Acta Electronica Sinica, vol. 39, no. 7, pp. 1563–1568, 2011. View at: Google Scholar
 R. Rubinstein, A. M. Bruckstein, and M. Elad, “Dictionaries for sparse representation modeling,” Proceedings of the IEEE, vol. 98, no. 6, pp. 1045–1057, 2010. View at: Publisher Site  Google Scholar
 Y. He, T. Gan, W. Chen, and H. Wang, “Multistage image denoising based on correlation coefficient matching and sparse dictionary pruning,” Signal Processing, vol. 92, no. 1, pp. 139–149, 2012. View at: Publisher Site  Google Scholar
 K. Labusch, E. Barth, and T. Martinetz, “Softcompetitive learning of sparse codes and its application to image reconstruction,” Neurocomputing, vol. 74, no. 9, pp. 1418–1428, 2011. View at: Publisher Site  Google Scholar
 Z. Xue, D. Han, and J. Tian, “Fast and robust reconstruction approach for sparse fluorescence tomography based on adaptive matching pursuit,” in Proceedings of the IEEE Conference on Asia Communications and Photonics, pp. 1–6, November 2011. View at: Publisher Site  Google Scholar
 D. Needell and R. Vershynin, “Signal recovery from incomplete and inaccurate measurements via regularized orthogonal matching pursuit,” IEEE Journal on Selected Topics in Signal Processing, vol. 4, no. 2, pp. 310–316, 2010. View at: Publisher Site  Google Scholar
 T. Blumensath and M. E. Davies, “Stagewise weak gradient pursuits,” IEEE Transactions on Signal Processing, vol. 57, no. 11, pp. 4333–4346, 2009. View at: Publisher Site  Google Scholar
 M. Aharon, M. Elad, and A. Bruckstein, “KSVD: an algorithm for designing overcomplete dictionaries for sparse representation,” IEEE Transactions on Signal Processing, vol. 54, no. 11, pp. 4311–4322, 2006. View at: Publisher Site  Google Scholar
 J. Mairal, G. Sapiro, and M. Elad, “Learning multiscale sparse representations for image and video restoration,” Multiscale Modeling and Simulation, vol. 7, no. 1, pp. 214–241, 2008. View at: Publisher Site  Google Scholar
 W.Q. Lim, “The discrete shearlet transform: a new directional transform and compactly supported shearlet frames,” IEEE Transactions on Image Processing, vol. 19, no. 5, pp. 1166–1180, 2010. View at: Publisher Site  Google Scholar
 T. Kanungo, D. M. Mount, N. S. Netanyahu, C. D. Piatko, R. Silverman, and A. Y. Wu, “An efficient kmeans clustering algorithms: analysis and implementation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, no. 7, pp. 881–892, 2002. View at: Publisher Site  Google Scholar
 I. Daubechies, M. Defrise, and C. de Mol, “An iterative thresholding algorithm for linear inverse problems with a sparsity constraint,” Communications on Pure and Applied Mathematics, vol. 57, no. 11, pp. 1413–1457, 2004. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 M. Poderico, S. Parrilli, G. Poggi, and L. Verdoliva, “Sigmoid shrinkage for BM3D denoising algorithm,” in Proceedings of the IEEE International Workshop on Multimedia Signal Processing (MMSP '10), pp. 423–426, October 2010. View at: Publisher Site  Google Scholar  Zentralblatt MATH
Copyright
Copyright © 2014 Wu Qidi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.