Convergent Plug-and-Play with Proximal Denoiser and Unconstrained Regularization Parameter

Journal of Mathematical Imaging and Vision(2024)

引用 0|浏览7
暂无评分
摘要
In this work, we present new proofs of convergence for plug-and-play (PnP) algorithms. PnP methods are efficient iterative algorithms for solving image inverse problems where regularization is performed by plugging a pre-trained denoiser in a proximal algorithm, such as Proximal Gradient Descent (PGD) or Douglas–Rachford splitting (DRS). Recent research has explored convergence by incorporating a denoiser that writes exactly as a proximal operator. However, in these works, the corresponding PnP algorithm has the drawback to be necessarily run with stepsize equal to 1. The stepsize condition for nonconvex convergence of the proximal algorithm in use then translates to restrictive conditions on the regularization parameter of the inverse problem. This can severely degrade the restoration capacity of the algorithm. In this paper, we present two remedies for this limitation. First, we provide a novel convergence proof for PnP-DRS that does not impose any restriction on the regularization parameter. Second, we examine a relaxed version of the PGD algorithm that converges across a broader range of regularization parameters. Our experimental study, conducted on deblurring and super-resolution experiments, demonstrate that these two solutions both enhance the accuracy of image restoration.
更多
查看译文
关键词
Nonconvex optimization,Inverse problems,Plug-and-play
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要