Issue 
Wuhan Univ. J. Nat. Sci.
Volume 28, Number 1, February 2023



Page(s)  53  60  
DOI  https://doi.org/10.1051/wujns/2023281053  
Published online  17 March 2023 
Computer Science
CLC number: TP 751
Group Sparsity Residual Constraint Image Denoising Model with 𝒍_{1}/𝒍_{2} Regularization
^{1}
Key Laboratory of Multidisciplinary Management and Control of Complex Systems of Anhui Higher Education Institutes, Anhui University of Technology, Maanshan 243002, Anhui, China
^{2}
School of Mathematics and Physics, Anhui University of Technology, Maanshan 243032, Anhui, China
^{†} To whom correspondence should be addressed. Email: zt9877@163.com
Received:
18
October
2022
Group sparse residual constraint with nonlocal priors (GSRC) has achieved great success in image restoration producing stateoftheart performance. In the GSRC model, the norm minimization is employed to reduce the group sparse residual. In recent years, nonconvex regularization terms have been widely used in image denoising problems, which have achieved better results in denoising than convex regularization terms. In this paper, we use the ratio of the and norm instead of the norm to propose a new image denoising model, i.e., a group sparse residual constraint model with minimization (GSRC). Due to the computational difficulties arisen from the nonconvexity and nonlinearity, we focus on a constrained optimization problem that can be solved by alternative direction method of multipliers (ADMM). Experimental results of image denoising show that the proposed model outperforms several stateoftheart image denoising methods both visually and quantitatively.
Key words: image denoising / l_{1}/l_{2} minimization / group sparse representation
Biography: WU Di, male, Master candidate, research direction: image processing. Email: wudi2020wudi@163.com
Fundation item: Supported by the Open Fund of Key Laboratory of Anhui Higher Education Institutes (CS202107) , the National Natural Science Foundation of China (61701004) , the Outstanding Young Talents Support Program of Anhui Province (GXYQ 2021178 ) and University Natural Science Research Project of Anhui Province of China ( KJ2020A0238)
© Wuhan University 2023
This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
0 Introduction
The goal of image denoising is to estimate the latent image from its degraded observation , which is typically an illposed inverse problem. To tackle the illposed nature of this problem, regularization methods based on various image prior information have to be incorporated into the image denoising process, which is usually realized by minimizing the objective function as
where and represent the original image and the degraded image, respectively. The first term is the data fidelity and the second term is the regularization term. is a regularization parameter that balances these two terms. During the past decades, a variety of image denoising methods have been proposed, such as total variation^{[1,2]}, sparse representation^{[35]}, nonlocal selfsimilarity^{[6,7]}, and neural network models^{[810]}.
Nowadays, sparse representation has become a mainstream direction in image processing. Aharon et al^{[11] }used singular value decomposition to update the basic atoms and corresponding coefficients in a minimum error criterion, and pioneered the KSVD (singular value decomposition) algorithm with a Ktime optimization process in 2006. In dictionary learning and sparse coding, each block is treated as an independent individual, while blocks are generally correlated with each other, which is usually ignored. Mairal et al^{[12] }assumed that the sparse coefficients of similar blocks had the same support mechanism and combined the image redundancy information to improve the KSVD denoising performance by learning simultaneous sparse coding (LSSC). Dong et al^{[13]} trained principal component analysis (PCA) dictionaries by clustering all image blocks, and for each patch group, the most appropriate local PCA dictionary was selected to encode it, thus called the nonlocal centralized sparse representation (NCSR) model. This NCSR method introduces the concept of "coding residuals". Zhang et al^{[14] }formally proposed the concept of "group" as the basic processing unit of sparse coding, and proposed a groupbased sparse representation (GSR) model to extend the mutually independent patch sparse to group sparse which is applicable to both nonlocal similarity and local sparsity of images. Zha et al^{[15]} proposed a group sparse residual constraint model with nonlocal priors (GSRCNLP) and used soft thresholding to achieve image denoising, which essentially combines the NCSR method and the GSR model to upgrade the patchbased sparse coding residuals to the groupbased sparse residuals.
Other than the popular norm, there are many regularization functions to promote sparsity. Bai et al^{[16] }proposed an adaptive correction term to alleviate the over constraint of the norm . The regularization term is also used to improve sparsity and can alleviate the overconstraint of norm to some extent ^{[17]}. In addition to this, many nonconvex regularization terms have been proposed, such as norm ^{[18]}, norm ^{[19]}, and Logarithmic regularization term^{[20]}.
In this paper, we study the ratio of the and norms, denoted as , to promote the sparsity of group sparse residual. The contributions of the paper are summarized as follows:
First, the regularization term is used to reduce the group sparse residual, and then a group sparse residual constraint model with minimization is proposed. The proposed model is successfully solved by the alternative direction method of multipliers (ADMM) algorithm.
Second, experimental results on image denoising show that the proposed scheme is feasible and outperforms many stateoftheart methods in both objective and perceptual quality.
1 Background and Related Work
1.1 Group Sparse Representation
In the framework of GSR method^{[14]}, an image with size is divided into overlapped patches of size , and each patch is denoted by a vector . Then for each patch , similar matched patches are selected from a sized searching window by Knearest neighbor (KNN) algorithm^{[21]}. All patches in set are stacked into a matrix , i.e., . This matrix consisting of patches with similar structures is called a patch group, where denotes the patch in the patch group.
Give a dictionary , which is often learned from each group, such as PCA dictionary^{[22]}. Each group can be sparsely represented by solving the following norm minimization problem:
where represents the group sparse coefficient of each group , i.e., . is an adaptive PCA dictionary learned from each patch group.
1.2 Construction of Group Sparse Coefficient Residuals
In Ref.[18], the authors pointed out that traditional GSR models fail to faithfully estimate the sparsity of each group due to the degradation of the observed image, and they showed that the quality of image restoration largely depends on the group sparsity residual, which is defined as the difference between each degraded group sparse coefficient and the corresponding original group sparse coefficient :
So, they proposed the group sparsity residual constraint model to boost up the accuracy of group sparse coefficient , which can be written as
where is the group sparse coefficient of each group corresponding to the original image. As the original image is not available in image denoising, the nonlocal selfsimilarity prior is used to estimate . For each group including nonlocal similar patches, a good estimation of can be computed by the weighted average of each vector in . We have
where , and represent the and vector of and , respectively. After this, is copied by times to estimate . We can obtain . In Eq. (5), is set to be inversely proportional to the distance between the target patch and its similar patch , i.e.,
where is a predefined constant and is a normalization factor ^{[23]}.
1.3 Minimization
In Ref.[17], the authors employed the norm to promote the sparsity and proposed the following model for sparse recovery:
Two main advantages of minimization are scale invariant and parameter free. As the ratio of and is nonconvex and nonlinear, solving model (6) and analyzing the convergence become theoretically difficult. The authors of Ref.[17] applied ADMM for minimizing model (11), and verified the convergence empirically by plotting residual errors and objective functions. In Ref.[17], the authors applied the norm on the image gradient for MRI reconstruction problem.
2 The GSRC Model for Image Denoising
In this subsection, we employ the ratio of and norm as the regularizer to promote the sparsity of the group sparse residual mentioned before. The proposed model is
where represents the dictionary and is the sparse coefficient.
As we mentioned before, patch group is the unite of the proposed sparse representation. Supposing , the corresponding patch group obtained by block matching is denoted by . Let , then Eq. (7) can be rewritten as
By means of Theorem I in Ref.[15], the following equation holds with a very large probability at each iteration.
Based on Eq. (9), minimization problem (8) is equivalent to
where , . Let and , is a PCA based dictionary learned from each group , so it is orthogonal. By utilizing ADMM and introducing two auxiliary variables and , Eq. (10) can be transformed into another equivalent constraint form:
The augmented Lagrange function plays a central role in the ADMM method. For the problem (11), the augmented Lagrange function can be written as
Then, the ADMM iteration goes as follows
Next, we will discuss the optimization strategies for solving each subproblem one by one.
2.1 SubProblem
The subproblem of is a least square problem. From the optimal conditions, we can obtain
The solution of is
2.2 SubProblem
The subproblem is equivalent to
It has a closed form solution via soft shrinkage, i.e.,
where represents pointwise product.
2.3 SubProblem
As for the subproblem, let , and the minimization problem reduces to
If , then any vector with is a solution to the minimization problem. If , then is the solution. If , by taking derivative of the objective function with respect to , we obtain
As a result, there exists a positive number such that . Thus, Eq. (21) reduces to
Given , we denote
Then is the root of
Let , we can find has only one real root, which has the following closedform:
where .
In summary, we have
where is a random vector with the norm to be .
Finally, the final estimated group patch is computed by . We obtain the complete image by putting the group back to its original position and averaging the overlapping pixels.
2.4 Settings of Parameters and Iterations
In addition, to obtain better results, we can perform the following several iterations of the denoising process. In the iteration, an iterative regularization strategy is used to update the estimation of the noise variance, and therefore updating . The standard deviation of the noise in the iteration is adjusted to
where is a constant.
Since the noise variance changes after each regularization process, the parameter , which is used to balance the fidelity and regularization terms, needs to change along with it. After each regularization, the parameter of group is set to
where denotes the estimated standard deviation of and are constants. The complete description of the proposed GSRC model for image denoising employing ADMM framework is presented in Algorithm 1.
Algorithm 1 The proposed GSRCl_{1}/l_{2} model for image denoising 

1. Require: Noisy image . 2. Initialization: Input , , , , , , , , , , ,, , , , and . 3. For to MaxIter 4. Iterative Regularization 5. For each patch in do 6. Find nonlocal similar patches from group . 7. Obtain dictionary by using PCA. 8. Update by computing . 9. Update by computing Eq. (5). 10. Update by computing Eq. (23). 11. While < ADMMIter or 12. Update by computing Eq. (18). 13. Update by computing Eq. (19). 14. Compute by computing Eq. (22). 15. Update via (16). 16. Update via (17). 17. Get the estimation: . 18. End while 19. End for 20. Aggregate to the denoised image . 21. End for 22. Output: The final denoised image . 
3 Experiments
In this section, we will present extensive experimental results to evaluate the denoising performance of the proposed GSRC model. To evaluate the quality of the recovered images, the peak signal to noise ratio (PSNR) and structural similarity (SSIM) metrics are used. The source codes of all comparison methods were obtained from the original authors, and we used the default parameter settings. For color images, image restoration operators are only performed for the luminance component in this paper.
3.1 Setting Adaptive Parameters of the GSRC Model
In this subsection, we report the performance of the proposed GSRC for image denoising and compare them with some leading denoising methods, including BM3D^{[24]}, NCSR^{[13]}, PGPD^{[25]}, SAIST^{[26] }and GSRCNLP ^{[15]}. The parameter setting of the proposed GSRC model for image denoising is as follows. The size of searching window for block matching is set to . All experiments are conducted under the MATLAB 2018a environment on a computer with Intel (R) Core (TM) i710750H with 2.60 GHz CPU and NVIDIA GeForce GTX 1650 Ti and Windows 10 platform. In the GSRC model, the size of patch is set to for . In addition, the parameter settings of the model are different. The parameters are (0.6,0.1,0.023,0.001), (0.7,0.1,0.003 4,0.001) and (0.7,0.1,0.0034,0.001) set to for , and .
3.2 Analysis of Results
Note that BM3D, PGPD and SAIST are the GSRbased or NSSbased image denoising methods. NCSR and GSRCNLP are group sparsity residual based image denoising methods. We test the denoising performance on the presented test images and the highest PSNR is recorded in this paper. The PSNR results for all methods are shown in Table 1.
The proposed GSRC has an average PSNR increase of 0.26dB, 0.32dB, 0.27dB, 0.07dB and 0.03dB compared with BM3D, NCSR, PGPD, SAIST and GSRCNLP, respectively. In addition, the proposed GSRC has an average SSIM increase of 0.013 3, 0.012 9, 0.014 4, 0.007 2 and 0.003 3 compared with BM3D, NCSR, PGPD, SAIST and GSRCNLP, respectively. We can observe that the proposed GSRC can obtain better results than other competing methods.
Figures 1 and 2 show the visual comparison of the images Barbara and Mural, respectively. To better compare the details of the reconstructed images, we zoom in the same parts of each image to twice the original size. It can be seen that BM3D, NCSR, SAIST and GSRCNLP are prone to over smooth the images. PGPD does not protect image details. In contrast, the proposed GSRC is able to preserve local structure of the image more effectively than other competing approaches.
Fig. 1 Denoising results on image Barbaraby different methods (noise level )
Denoising results on image Barbaraby different methods (noise level ) 
Fig. 2 Denoising results on image Mural by different methods (noise level )
Denoising results on image Mural by different methods (noise level ) 
PSNR and SSIM result by denoising methods
3.3 Empirical Convergence Analysis of GSRC Model
The existing literature on the ADMM convergence requires the existence of one separable function in the objective function, whose gradient is Lipschitz continuous. However, the proposed GSRC model does not satisfy this assumption. Therefore, it is difficult to analyze the convergence theoretically. Instead, in this subsection we will show the convergence empirically by plotting residual errors and objective functions, which provides strong supports for the convergence validation. In Fig. 3, we empirically demonstrate the convergence of the proposed GSRC algorithm. There are two auxiliary variables and in , i.e., and .Figure 3(a) shows the values of and with respect to iteration counter . Figure 3(b) shows the values of the objective function (11) versus iteration counter . The constraints and objective functions in Fig. 3 decrease rapidly with respect to the iteration counters, which is the heuristic evidence of the convergence of the Algorithm 1.
Fig. 3 Plots of residual errors and objective function for empirically demonstrating the convergence
Plots of residual errors and objective function for empirically demonstrating the convergence 
4 Conclusion
To improve the performance of group sparse based image restoration approaches, we introduce an effective sparsity promoting strategies, i.e., minimization to promote the sparsity of the group sparse residual. We use the nonlocal selfsimilar prior of images alone with a selfsupervised learning scheme to obtain good estimates of the group sparsity coefficients for each original group. Then we adapt the minimization to reduce the group sparse residual. Compared with minimization in GSRCNLP method, the proposed GSRC can lead to better performance in image denoising. We apply the ADMM algorithm to solve the models and explore the convergence of the proposed algorithm. Experimental results show that the model is comparable with the tested denoising methods, and outperforms many stateoftheart image denoising methods in terms of both objective and perceptual quality metrics.
References
 BlomgrenP, ChanT F. Color TV: Total variation methods for restoration of vectorvalued images[J]. IEEE Transactions on Image Processing, 1998, 7(3):304309. [NASA ADS] [CrossRef] [PubMed] [Google Scholar]
 BertalmioM, CasellesV, RougeB, et al. A TV based restoration model with local constraints. [J]. Journal of Scientific Computing, 2003, 19(1):95122. [Google Scholar]
 LiH, LiuF. Image denoising via sparse and redundant representations over learned dictionaries in wavelet domain[C]// 2009 Fifth International Conference on Image and Graphics. Washington D C: IEEE, 2009: 754758. [Google Scholar]
 MairalJ, EladM, SapiroG. Sparse representation for color image restoration[J]. IEEE Transactions on Image Processing, 2007, 17(1):5369. [Google Scholar]
 AharonM, EladM, BrucksteinA M. On the uniqueness of overcomplete dictionaries, and a practical way to retrieve them[J]. Linear Algebra & Its Applications, 2006, 416(1):4867. [Google Scholar]
 ZhaZ, YuanX, ZhuC, et al. Image restoration via simultaneous nonlocal selfsimilarity priors[J]. IEEE Transactions on Image Processing, 2020, 29: 85618576. [NASA ADS] [CrossRef] [Google Scholar]
 GuS H, ZhangL, ZuoW M, et al. Weighted nuclear norm minimization with application to image denoising[C]// 2014 IEEE Conference on Computer Vision and Pattern Recognition. Washington D C : IEEE, 2014: 28622869. [Google Scholar]
 ZhangK, ZuoW M, ChenY J, et al. Beyond a Gaussian denoiser: Residual learning of deep CNN for image denoising[J]. IEEE Transactions on Image Processing, 2016, 26(7):31423155. [Google Scholar]
 DongW S, WangP Y, YinW T, et al. Denoising prior driven deep neural network for image restoration[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019, 41(10): 23052318. [CrossRef] [PubMed] [Google Scholar]
 ZhangK, ZuoW M, GuS H, et al. Learning deep CNN denoiser prior for image restoration[C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Washington D C : IEEE, 2017: 28082817. [Google Scholar]
 AharonM, EladM, Bruckstein. A. KSVD: An algorithm for designing overcomplete dictionaries for sparse representation[J]. IEEE Transactions on Signal Processing, 2006, 54(11): 43114322. [CrossRef] [Google Scholar]
 MairalJ, BachF, PonceJ, et al. Nonlocal sparse models for image restoration[C]// 2009 IEEE 12th International Conference on Computer Vision. Washington D C : IEEE, 2009: 22722279 [Google Scholar]
 DongW S, ZhangL, ShiG M, et al. Nonlocally centralized sparse representation for image restoration[J]. IEEE Transcation on Image Processing, 2013, 22(4):16201630. [NASA ADS] [CrossRef] [PubMed] [Google Scholar]
 ZhangJ, ZhaoD B, GaoW. Groupbased sparse representation for image restoration[J]. IEEE Transactions on Image Processing, 2014, 23(8): 33363351. [NASA ADS] [CrossRef] [MathSciNet] [PubMed] [Google Scholar]
 ZhaZ Y, YuanX, WenB H, et al. Group sparsity residual constraint with nonlocal priors for image restoration[J]. IEEE Transactions on Image Processing, 2020, 29: 89608975. [NASA ADS] [CrossRef] [MathSciNet] [Google Scholar]
 BaiM, ZhangX, ShaoQ. Adaptive correction procedure for TVL1 image deblurring under impulse noise[J]. Inverse Problems, 2016, 32(8): 289338. [Google Scholar]
 RahimiY, WangC, DongH B, et al. A scaleinvariant approach for sparse signal recovery[J]. SIAM Journal on Scientific Computing, 2019, 41(6):A3649A3672. [Google Scholar]
 XuZ B, ChangX Y, XuF M, et al. L1/2 regularization: A thresholding representation theory and a fast solver[J]. IEEE Transactions on Neural Networks and Learning Systems, 2012, 23(7):10131027. [CrossRef] [PubMed] [Google Scholar]
 ZuoW M, MengD Y, ZhangL, et al. A generalized iterated shrinkage algorithm for nonconvex sparse coding[C]// 2013 IEEE International Conference on Computer Vision. Washington D C: IEEE, 2013: 217224. [Google Scholar]
 JiaX, FengX. Bayesian inference for adaptive low rank and sparse matrix estimation[J]. Neurocomputing, 2018(5):7183. [Google Scholar]
 KellerJ M, GrayM R, GivensJ A. A fuzzy knearest neighbor algorithm[J]. IEEE Transactions on Systems Man & Cybernetics,1985, SMC15(4): 580585. [Google Scholar]
 DongW S, LeiZ, ShiG M, et al. Image deblurring and superresolution by adaptive sparse domain selection and adaptive regularization[J]. IEEE Transactions on Image Processing, 2011, 20(7):18381857. [CrossRef] [MathSciNet] [PubMed] [Google Scholar]
 BuadesA, CollB, MorelJ M. A nonlocal algorithm for image denoising[C]// 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05). Washington D C: IEEE, 2005, 2: 6065. [Google Scholar]
 DabovK, FoiA, KatkovnikV, et al. Image denoising by sparse 3D transformdomain collaborative filtering[J]. IEEE Transactions on Image Processing, 2007, 16(8):20802095. [CrossRef] [Google Scholar]
 XuJ, ZhangL , ZuoW M, et al. Patch group based nonlocal selfsimilarity prior learning for image denoising[C]// 2015 IEEE International Conference on Computer Vision (ICCV). Washington D C: IEEE, 2016: 244252. [Google Scholar]
 DongW S, ShiG M, LiX. Nonlocal image restoration with bilateral variance estimation: A lowrank approach[J]. IEEE Transactions on Image Processing, 2013, 22(2):700711. [NASA ADS] [CrossRef] [MathSciNet] [PubMed] [Google Scholar]
All Tables
All Figures
Fig. 1 Denoising results on image Barbaraby different methods (noise level )
Denoising results on image Barbaraby different methods (noise level ) 

In the text 
Fig. 2 Denoising results on image Mural by different methods (noise level )
Denoising results on image Mural by different methods (noise level ) 

In the text 
Fig. 3 Plots of residual errors and objective function for empirically demonstrating the convergence
Plots of residual errors and objective function for empirically demonstrating the convergence 

In the text 
Current usage metrics show cumulative count of Article Views (fulltext article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 4896 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.