Open Access
Issue
Wuhan Univ. J. Nat. Sci.
Volume 28, Number 1, February 2023
Page(s) 53 - 60
DOI https://doi.org/10.1051/wujns/2023281053
Published online 17 March 2023

© Wuhan University 2023

Licence Creative CommonsThis is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

0 Introduction

The goal of image denoising is to estimate the latent image from its degraded observation , which is typically an ill-posed inverse problem. To tackle the ill-posed nature of this problem, regularization methods based on various image prior information have to be incorporated into the image denoising process, which is usually realized by minimizing the objective function as

(1)

where and represent the original image and the degraded image, respectively. The first term is the data fidelity and the second term is the regularization term. is a regularization parameter that balances these two terms. During the past decades, a variety of image denoising methods have been proposed, such as total variation[1,2], sparse representation[3-5], non-local self-similarity[6,7], and neural network models[8-10].

Nowadays, sparse representation has become a mainstream direction in image processing. Aharon et al[11] used singular value decomposition to update the basic atoms and corresponding coefficients in a minimum error criterion, and pioneered the K-SVD (singular value decomposition) algorithm with a K-time optimization process in 2006. In dictionary learning and sparse coding, each block is treated as an independent individual, while blocks are generally correlated with each other, which is usually ignored. Mairal et al[12] assumed that the sparse coefficients of similar blocks had the same support mechanism and combined the image redundancy information to improve the K-SVD denoising performance by learning simultaneous sparse coding (LSSC). Dong et al[13] trained principal component analysis (PCA) dictionaries by clustering all image blocks, and for each patch group, the most appropriate local PCA dictionary was selected to encode it, thus called the non-local centralized sparse representation (NCSR) model. This NCSR method introduces the concept of "coding residuals". Zhang et al[14] formally proposed the concept of "group" as the basic processing unit of sparse coding, and proposed a group-based sparse representation (GSR) model to extend the mutually independent patch sparse to group sparse which is applicable to both non-local similarity and local sparsity of images. Zha et al[15] proposed a group sparse residual constraint model with non-local priors (GSRC-NLP) and used soft thresholding to achieve image denoising, which essentially combines the NCSR method and the GSR model to upgrade the patch-based sparse coding residuals to the group-based sparse residuals.

Other than the popular norm, there are many regularization functions to promote sparsity. Bai et al[16] proposed an adaptive correction term to alleviate the over constraint of the norm . The regularization term is also used to improve sparsity and can alleviate the over-constraint of norm to some extent [17]. In addition to this, many non-convex regularization terms have been proposed, such as norm [18], norm [19], and Logarithmic regularization term[20].

In this paper, we study the ratio of the and norms, denoted as , to promote the sparsity of group sparse residual. The contributions of the paper are summarized as follows:

First, the regularization term is used to reduce the group sparse residual, and then a group sparse residual constraint model with minimization is proposed. The proposed model is successfully solved by the alternative direction method of multipliers (ADMM) algorithm.

Second, experimental results on image denoising show that the proposed scheme is feasible and outperforms many state-of-the-art methods in both objective and perceptual quality.

1 Background and Related Work

1.1 Group Sparse Representation

In the framework of GSR method[14], an image with size is divided into overlapped patches of size , and each patch is denoted by a vector . Then for each patch , similar matched patches are selected from a sized searching window by K-nearest neighbor (KNN) algorithm[21]. All patches in set are stacked into a matrix , i.e., . This matrix consisting of patches with similar structures is called a patch group, where denotes the patch in the patch group.

Give a dictionary , which is often learned from each group, such as PCA dictionary[22]. Each group can be sparsely represented by solving the following norm minimization problem:

(2)

where represents the group sparse coefficient of each group , i.e., . is an adaptive PCA dictionary learned from each patch group.

1.2 Construction of Group Sparse Coefficient Residuals

In Ref.[18], the authors pointed out that traditional GSR models fail to faithfully estimate the sparsity of each group due to the degradation of the observed image, and they showed that the quality of image restoration largely depends on the group sparsity residual, which is defined as the difference between each degraded group sparse coefficient and the corresponding original group sparse coefficient :

(3)

So, they proposed the group sparsity residual constraint model to boost up the accuracy of group sparse coefficient , which can be written as

(4)

where is the group sparse coefficient of each group corresponding to the original image. As the original image is not available in image denoising, the non-local self-similarity prior is used to estimate . For each group including non-local similar patches, a good estimation of can be computed by the weighted average of each vector in . We have

(5)

where , and represent the and vector of and , respectively. After this, is copied by times to estimate . We can obtain . In Eq. (5), is set to be inversely proportional to the distance between the target patch and its similar patch , i.e.,

where is a predefined constant and is a normalization factor [23].

1.3 Minimization

In Ref.[17], the authors employed the norm to promote the sparsity and proposed the following model for sparse recovery:

(6)

Two main advantages of minimization are scale invariant and parameter free. As the ratio of and is non-convex and non-linear, solving model (6) and analyzing the convergence become theoretically difficult. The authors of Ref.[17] applied ADMM for minimizing model (11), and verified the convergence empirically by plotting residual errors and objective functions. In Ref.[17], the authors applied the norm on the image gradient for MRI reconstruction problem.

2 The GSRC- Model for Image Denoising

In this subsection, we employ the ratio of and norm as the regularizer to promote the sparsity of the group sparse residual mentioned before. The proposed model is

(7)

where represents the dictionary and is the sparse coefficient.

As we mentioned before, patch group is the unite of the proposed sparse representation. Supposing , the corresponding patch group obtained by block matching is denoted by . Let , then Eq. (7) can be rewritten as

(8)

By means of Theorem I in Ref.[15], the following equation holds with a very large probability at each iteration.

(9)

Based on Eq. (9), minimization problem (8) is equivalent to

(10)

where , . Let and , is a PCA based dictionary learned from each group , so it is orthogonal. By utilizing ADMM and introducing two auxiliary variables and , Eq. (10) can be transformed into another equivalent constraint form:

(11)

The augmented Lagrange function plays a central role in the ADMM method. For the problem (11), the augmented Lagrange function can be written as

(12)

Then, the ADMM iteration goes as follows

(13)

(14)

(15)

(16)

(17)

Next, we will discuss the optimization strategies for solving each sub-problem one by one.

2.1 Sub-Problem

The sub-problem of is a least square problem. From the optimal conditions, we can obtain

The solution of is

(18)

2.2 Sub-Problem

The sub-problem is equivalent to

It has a closed form solution via soft shrinkage, i.e.,

(19)

where represents point-wise product.

2.3 Sub-Problem

As for the sub-problem, let , and the minimization problem reduces to

(20)

If , then any vector with is a solution to the minimization problem. If , then is the solution. If , by taking derivative of the objective function with respect to , we obtain

(21)

As a result, there exists a positive number such that . Thus, Eq. (21) reduces to

Given , we denote

Then is the root of

Let , we can find has only one real root, which has the following closed-form:

where .

In summary, we have

(22)

where is a random vector with the norm to be .

Finally, the final estimated group patch is computed by . We obtain the complete image by putting the group back to its original position and averaging the overlapping pixels.

2.4 Settings of Parameters and Iterations

In addition, to obtain better results, we can perform the following several iterations of the denoising process. In the iteration, an iterative regularization strategy is used to update the estimation of the noise variance, and therefore updating . The standard deviation of the noise in the iteration is adjusted to

where is a constant.

Since the noise variance changes after each regularization process, the parameter , which is used to balance the fidelity and regularization terms, needs to change along with it. After each regularization, the parameter of group is set to

(23)

where denotes the estimated standard deviation of and are constants. The complete description of the proposed GSRC- model for image denoising employing ADMM framework is presented in Algorithm 1.

Algorithm 1 The proposed GSRC-l1/l2 model for image denoising
1. Require: Noisy image .
2. Initialization: Input , , , , , , , , , , ,, , , , and .
3. For to Max-Iter
4.  Iterative Regularization
5.  For each patch in do
6.   Find non-local similar patches from group .
7.   Obtain dictionary by using PCA.
8.   Update by computing .
9.   Update by computing Eq. (5).
10.  Update by computing Eq. (23).
11.  While < ADMM-Iter or
12.    Update by computing Eq. (18).
13.    Update by computing Eq. (19).
14.    Compute by computing Eq. (22).
15.    Update via (16).
16.    Update via (17).
17.    Get the estimation: .
18.   End while
19.  End for
20.  Aggregate to the denoised image .
21. End for
22. Output: The final denoised image .

3 Experiments

In this section, we will present extensive experimental results to evaluate the denoising performance of the proposed GSRC- model. To evaluate the quality of the recovered images, the peak signal to noise ratio (PSNR) and structural similarity (SSIM) metrics are used. The source codes of all comparison methods were obtained from the original authors, and we used the default parameter settings. For color images, image restoration operators are only performed for the luminance component in this paper.

3.1 Setting Adaptive Parameters of the GSRC- Model

In this subsection, we report the performance of the proposed GSRC- for image denoising and compare them with some leading denoising methods, including BM3D[24], NCSR[13], PGPD[25], SAIST[26] and GSRC-NLP [15]. The parameter setting of the proposed GSRC- model for image denoising is as follows. The size of searching window for block matching is set to . All experiments are conducted under the MATLAB 2018a environment on a computer with Intel (R) Core (TM) i7-10750H with 2.60 GHz CPU and NVIDIA GeForce GTX 1650 Ti and Windows 10 platform. In the GSRC- model, the size of patch is set to for . In addition, the parameter settings of the model are different. The parameters are (0.6,0.1,0.023,0.001), (0.7,0.1,0.003 4,0.001) and (0.7,0.1,0.0034,0.001) set to for , and .

3.2 Analysis of Results

Note that BM3D, PGPD and SAIST are the GSR-based or NSS-based image denoising methods. NCSR and GSRC-NLP are group sparsity residual based image denoising methods. We test the denoising performance on the presented test images and the highest PSNR is recorded in this paper. The PSNR results for all methods are shown in Table 1.

The proposed GSRC- has an average PSNR increase of 0.26dB, 0.32dB, 0.27dB, 0.07dB and 0.03dB compared with BM3D, NCSR, PGPD, SAIST and GSRC-NLP, respectively. In addition, the proposed GSRC- has an average SSIM increase of 0.013 3, 0.012 9, 0.014 4, 0.007 2 and 0.003 3 compared with BM3D, NCSR, PGPD, SAIST and GSRC-NLP, respectively. We can observe that the proposed GSRC- can obtain better results than other competing methods.

Figures 1 and 2 show the visual comparison of the images Barbara and Mural, respectively. To better compare the details of the reconstructed images, we zoom in the same parts of each image to twice the original size. It can be seen that BM3D, NCSR, SAIST and GSRC-NLP are prone to over smooth the images. PGPD does not protect image details. In contrast, the proposed GSRC- is able to preserve local structure of the image more effectively than other competing approaches.

thumbnail Fig. 1 Denoising results on image Barbaraby different methods (noise level )

Denoising results on image Barbaraby different methods (noise level )

thumbnail Fig. 2 Denoising results on image Mural by different methods (noise level )

Denoising results on image Mural by different methods (noise level )

Table 1

PSNR and SSIM result by denoising methods

3.3 Empirical Convergence Analysis of GSRC- Model

The existing literature on the ADMM convergence requires the existence of one separable function in the objective function, whose gradient is Lipschitz continuous. However, the proposed GSRC- model does not satisfy this assumption. Therefore, it is difficult to analyze the convergence theoretically. Instead, in this subsection we will show the convergence empirically by plotting residual errors and objective functions, which provides strong supports for the convergence validation. In Fig. 3, we empirically demonstrate the convergence of the proposed GSRC- algorithm. There are two auxiliary variables and in , i.e., and .Figure 3(a) shows the values of and with respect to iteration counter . Figure 3(b) shows the values of the objective function (11) versus iteration counter . The constraints and objective functions in Fig. 3 decrease rapidly with respect to the iteration counters, which is the heuristic evidence of the convergence of the Algorithm 1.

thumbnail Fig. 3 Plots of residual errors and objective function for empirically demonstrating the convergence

Plots of residual errors and objective function for empirically demonstrating the convergence

4 Conclusion

To improve the performance of group sparse based image restoration approaches, we introduce an effective sparsity promoting strategies, i.e., minimization to promote the sparsity of the group sparse residual. We use the non-local self-similar prior of images alone with a self-supervised learning scheme to obtain good estimates of the group sparsity coefficients for each original group. Then we adapt the minimization to reduce the group sparse residual. Compared with minimization in GSRC-NLP method, the proposed GSRC- can lead to better performance in image denoising. We apply the ADMM algorithm to solve the models and explore the convergence of the proposed algorithm. Experimental results show that the model is comparable with the tested denoising methods, and outperforms many state-of-the-art image denoising methods in terms of both objective and perceptual quality metrics.

References

  1. BlomgrenP, ChanT F. Color TV: Total variation methods for restoration of vector-valued images[J]. IEEE Transactions on Image Processing, 1998, 7(3):304-309. [NASA ADS] [CrossRef] [PubMed] [Google Scholar]
  2. BertalmioM, CasellesV, RougeB, et al. A TV based restoration model with local constraints. [J]. Journal of Scientific Computing, 2003, 19(1):95-122. [Google Scholar]
  3. LiH, LiuF. Image denoising via sparse and redundant representations over learned dictionaries in wavelet domain[C]// 2009 Fifth International Conference on Image and Graphics. Washington D C: IEEE, 2009: 754-758. [Google Scholar]
  4. MairalJ, EladM, SapiroG. Sparse representation for color image restoration[J]. IEEE Transactions on Image Processing, 2007, 17(1):53-69. [Google Scholar]
  5. AharonM, EladM, BrucksteinA M. On the uniqueness of overcomplete dictionaries, and a practical way to retrieve them[J]. Linear Algebra & Its Applications, 2006, 416(1):48-67. [Google Scholar]
  6. ZhaZ, YuanX, ZhuC, et al. Image restoration via simultaneous nonlocal selfsimilarity priors[J]. IEEE Transactions on Image Processing, 2020, 29: 8561-8576. [NASA ADS] [CrossRef] [Google Scholar]
  7. GuS H, ZhangL, ZuoW M, et al. Weighted nuclear norm minimization with application to image denoising[C]// 2014 IEEE Conference on Computer Vision and Pattern Recognition. Washington D C : IEEE, 2014: 2862-2869. [Google Scholar]
  8. ZhangK, ZuoW M, ChenY J, et al. Beyond a Gaussian denoiser: Residual learning of deep CNN for image denoising[J]. IEEE Transactions on Image Processing, 2016, 26(7):3142-3155. [Google Scholar]
  9. DongW S, WangP Y, YinW T, et al. Denoising prior driven deep neural network for image restoration[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019, 41(10): 2305-2318. [CrossRef] [PubMed] [Google Scholar]
  10. ZhangK, ZuoW M, GuS H, et al. Learning deep CNN denoiser prior for image restoration[C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Washington D C : IEEE, 2017: 2808-2817. [Google Scholar]
  11. AharonM, EladM, Bruckstein. A. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation[J]. IEEE Transactions on Signal Processing, 2006, 54(11): 4311-4322. [CrossRef] [Google Scholar]
  12. MairalJ, BachF, PonceJ, et al. Non-local sparse models for image restoration[C]// 2009 IEEE 12th International Conference on Computer Vision. Washington D C : IEEE, 2009: 2272-2279 [Google Scholar]
  13. DongW S, ZhangL, ShiG M, et al. Nonlocally centralized sparse representation for image restoration[J]. IEEE Transcation on Image Processing, 2013, 22(4):1620-1630. [NASA ADS] [CrossRef] [PubMed] [Google Scholar]
  14. ZhangJ, ZhaoD B, GaoW. Group-based sparse representation for image restoration[J]. IEEE Transactions on Image Processing, 2014, 23(8): 3336-3351. [NASA ADS] [CrossRef] [MathSciNet] [PubMed] [Google Scholar]
  15. ZhaZ Y, YuanX, WenB H, et al. Group sparsity residual constraint with non-local priors for image restoration[J]. IEEE Transactions on Image Processing, 2020, 29: 8960-8975. [NASA ADS] [CrossRef] [MathSciNet] [Google Scholar]
  16. BaiM, ZhangX, ShaoQ. Adaptive correction procedure for TVL1 image deblurring under impulse noise[J]. Inverse Problems, 2016, 32(8): 289-338. [Google Scholar]
  17. RahimiY, WangC, DongH B, et al. A scale-invariant approach for sparse signal recovery[J]. SIAM Journal on Scientific Computing, 2019, 41(6):A3649-A3672. [Google Scholar]
  18. XuZ B, ChangX Y, XuF M, et al. L-1/2 regularization: A thresholding representation theory and a fast solver[J]. IEEE Transactions on Neural Networks and Learning Systems, 2012, 23(7):1013-1027. [CrossRef] [PubMed] [Google Scholar]
  19. ZuoW M, MengD Y, ZhangL, et al. A generalized iterated shrinkage algorithm for non-convex sparse coding[C]// 2013 IEEE International Conference on Computer Vision. Washington D C: IEEE, 2013: 217-224. [Google Scholar]
  20. JiaX, FengX. Bayesian inference for adaptive low rank and sparse matrix estimation[J]. Neurocomputing, 2018(5):71-83. [Google Scholar]
  21. KellerJ M, GrayM R, GivensJ A. A fuzzy k-nearest neighbor algorithm[J]. IEEE Transactions on Systems Man & Cybernetics,1985, SMC-15(4): 580-585. [Google Scholar]
  22. DongW S, LeiZ, ShiG M, et al. Image deblurring and super-resolution by adaptive sparse domain selection and adaptive regularization[J]. IEEE Transactions on Image Processing, 2011, 20(7):1838-1857. [CrossRef] [MathSciNet] [PubMed] [Google Scholar]
  23. BuadesA, CollB, MorelJ M. A non-local algorithm for image denoising[C]// 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05). Washington D C: IEEE, 2005, 2: 60-65. [Google Scholar]
  24. DabovK, FoiA, KatkovnikV, et al. Image denoising by sparse 3-D transform-domain collaborative filtering[J]. IEEE Transactions on Image Processing, 2007, 16(8):2080-2095. [CrossRef] [Google Scholar]
  25. XuJ, ZhangL , ZuoW M, et al. Patch group based nonlocal self-similarity prior learning for image denoising[C]// 2015 IEEE International Conference on Computer Vision (ICCV). Washington D C: IEEE, 2016: 244-252. [Google Scholar]
  26. DongW S, ShiG M, LiX. Nonlocal image restoration with bilateral variance estimation: A low-rank approach[J]. IEEE Transactions on Image Processing, 2013, 22(2):700-711. [NASA ADS] [CrossRef] [MathSciNet] [PubMed] [Google Scholar]

All Tables

Table 1

PSNR and SSIM result by denoising methods

All Figures

thumbnail Fig. 1 Denoising results on image Barbaraby different methods (noise level )

Denoising results on image Barbaraby different methods (noise level )

In the text
thumbnail Fig. 2 Denoising results on image Mural by different methods (noise level )

Denoising results on image Mural by different methods (noise level )

In the text
thumbnail Fig. 3 Plots of residual errors and objective function for empirically demonstrating the convergence

Plots of residual errors and objective function for empirically demonstrating the convergence

In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.