Abstract

Nonlinear inverse problems arise in various fields ranging from scientific computation to engineering technology. Inverse problems are intrinsically ill-posed, and effective regularization techniques are necessary. The core of a suitable regularization method is to introduce the prior information of the model via an explicit or implicit regularization function. Plug-and-play regularization is a flexible framework that integrates the most effective denoising priors into an iterative algorithm, and it has recently shown great potential in the solution of linear ill-posed problems. Unlike traditional regularization methods, plug-and-play regularization does not require an explicit regularization function to represent the prior information of the model. In this work, by using total variation, block-matching and three-dimensional filtering, and fast and flexible denoising convolutional neural network denoisers, we propose a novel iterative regularization algorithm based on the alternating direction method of multipliers method. The combination of total variation and block-matching three-dimensional filtering regularizers can take advantage of the sparsity and nonlocal similarity in the solution of inverse problems. When combined with traditional and novel regularizers, deep neural networks have been shown to be an effective regularization approach, which can achieve state-of-the-art performance. Finally, we apply the proposed algorithm to the full waveform inversion problem to show the effectiveness of our method. Numerical results demonstrate that the proposed algorithm outperforms existing inversion methods in terms of quantitative measures and visual perceptual quality.

1. Introduction

A nonlinear inverse problem is generally expressed as the following least-squares optimization problem

(1)

where the model |$\bf {m}$| needs to be inverted from the measured data |$\bf {d}$|⁠. The forward modeling operator |$\mathcal {F(\bf {m})}$| is parametrized by |$\bf {m}$|⁠, which is governed by appropriate partial differential equations. Mathematically speaking, the problem (1) is usually nonconvex and severely ill-posed. Regularization methods are necessary to overcome the ill-posedness of the problem (1) and steer the overall procedure for optimization into an ideal direction (Tikhonov & Arsenin 1977).

After introducing regularization, problem (1) becomes

(2)

where the regularizer or regularization function |${\mathcal {R}(\bf {m})}$| includes prior knowledge about the expected solution |$\bf {m}$|⁠, and |$\lambda \gt 0$| is the regularization parameter (Wang 2016).

A popular choice for the regularization term |$\mathcal {R}(\bf {m})$| is the total variation (TV) function |$\Vert \nabla \bf {m}\Vert _{1}$|⁠, i.e., |$\mathcal {R}(\bf {m})$| is the |$l^{1}$|-norm of the model gradient (Rudin et al.1992). TV regularization assumes that the desired model is sparse in the derivative domain, and thus promotes piece-wise smooth solutions while preserving sharp discontinuances (Li & Han 2016). It is also very effective in full waveform inversion (FWI) problems, especially for large-velocity contrast structures, such as salt bodies embedded in sedimentary layers (Esser et al.2016; Lin & Huang 2014; Peng et al.2018). The TV norm is nondifferentiable and some proximal gradient-type algorithms should be used to solve the nonsmooth optimization problem (2) with |$\mathcal {R}(\bf {m})=\Vert \nabla \bf {m}\Vert _{1}$|⁠, where the proximal operator of |$\mathcal {R}$| is defined as

Note that the above proximal operator can be viewed as a Gaussian denoiser, which means it is possible to solve (1) with powerful regularizers embedded in excellent denoising methods such as block-matching three-dimensional (BM3D; Burger et al.2012; Dabov et al.2007) or deep learning denoisers (Sun et al.2019; Wu & Lin 2020; Zhang & Alkhalifah 2019). The core of the BM3D is to simultaneously exploit both the local sparsity of wavelet coefficients and the inherent nonlocal self-similarity properties of grouped image blocks. These two priors are very useful for image denoising, and the BM3D is often more effective compared with other existing denoisers.

The newest trend is deep learning, which has arisen as a promising mechanism and provides new ideas for solving ill-posed inverse problems. Deep-learning-based image denoisers have been developed rapidly, and it may be even possible to provide more performance gain than classical state-of-the-art denoisers such as TV or BM3D. For instance, Zhang et al. (2018) proposed a fast and flexible denoising convolutional neural network (FFDNet) by introducing a tunable noise level map as input. As a result, FFDNet exhibits excellent performance on both synthetic noisy images and real-world noisy images, and has flexibility in handling various types of noises.

With the mathematical equivalence between the proximal operator and denoiser, Venkatakrishnan et al. (2013) proposed plug-and-play (PnP) priors, and have received great success for regularized linear ill-posed problems (Cascarano et al.2022; Chan et al.2017; Rasti-Meymandi et al.2022). Plug-and-play is a highly flexible and effective framework for regularizing ill-posed problems by using more advanced denoisers within the alternating direction method of multipliers (ADMM; Boyd et al.2010) or other proximal-type algorithms. In this paper, we exploit the fact that different pieces of prior information could promote each other, and we further explore the effectiveness of PnP for solving the nonlinear ill-posed problem. More specifically, three popular denoisers (i.e., TV, BM3D, and FFDNet) are integrated into the ADMM algorithm by means of the PnP prior framework. Apparently, the proposed method takes full advantage of the sparsity (both global and local) and nonlocal self-similarity priors of the desired model, as well as the deep priors characterized by deep denoiser networks. The global sparsity is enforced by the TV norm for constructing the entire consistency and preserving edges, while the local sparsity is captured by the BM3D denoiser to preserve more features by exploiting model self-similarity prior. Such advantages suggest that there is the possibility of applying a powerful CNN denoiser to improve the ill-posedness of the inverse problem.

The paper is organized as follows. Section 2 depicts the proposed PnP regularization, which combines TV, BM3D, and FFDNet denoisers. Section 3 focuses on the applications to the FWI problem, and the experimental results and parameter selection are given. Section 4 presents the conclusion.

2. The proposed PnP regularization method

With the sparsity, nonlocal self-similarity, and deep priors all incorporated, problem (2) is formulated as a composite regularization problem

(3)

where |$\alpha$|⁠, |$\beta$|⁠, and |$\gamma$| are three regularization parameters, |$\Vert \cdot \Vert _{TV}=\Vert \nabla \bf {m}\Vert _{1}$| is the TV norm, and the last two items |$\Phi (\bf {m})$| and |$\mathcal {D}(\bf {m})$| are implicit regularizers that are deduced from the BM3D and FFDNet denoisers, respectively.

We set |$\mathcal {R}(\bf {m})=\alpha \Vert \bf {m}\Vert _{TV}+ \beta \Phi (\bf {m})+\gamma \mathcal {D}(\bf {m})$| in problem (3), by introducing an auxiliary variable |$\bf {v}$|⁠. Problem (3) can be converted into the following constrained optimization problem

(4)

Consider its augmented Lagrangian function

where |$\bf {u}$| is the Lagrange multiplier, and |$\rho$| is a penalty parameter.

To solve problem (4), we adopt the ADMM framework, which decomposes this optimization problem into simpler denoising subproblems, and then we solve the subproblems one by one. Algorithm 1 presents the pseudo code of our method, called plug-and-play regularization, or the PPR method. In step (a), the data fidelity term can be computed with any gradient-based optimization method including the steepest-descent method (Choi et al.2005) or quasi-Newton method, to obtain an initial value as a noisy solution. In steps (b)–(d), the variable |$\bf {v}$| can be split into three variables, |$\bf {v}_{1}$|⁠, |$\bf {v}_{2}$|⁠, and |$\bf {v}_{3}$|⁠, and the corresponding denoising operation is performed over each of them. In step (e), the Lagrange multiplier |$\bf {u}$| is updated. The convergence of the PPR method is guaranteed by the ADMM framework.

Algorithm 1

PPR method

Initialization:Pick initial guess |$\bf {u}_{0}$|⁠, |$\bf {v}^{0}$|⁠, parameters |$\rho$|⁠, |$\alpha$|⁠, |$\beta$|⁠, and |$\gamma$|⁠.
For |$k=0,1,\cdots ,K-1$|
(a) |$\bf {m}_{k+1}=\underset{\bf {m}}{\arg \min }\lbrace \frac{1}{2}\Vert \mathcal {F}(\bf {m})-\bf {d}\Vert _{2}^{2}+ {\frac{\rho }{2}\Vert {\bf {m}-\bf {v}^{k}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace ;$|
(b) |$\bf {v}_{1}=\underset{\bf {v}}{\arg \min }\lbrace \alpha \Vert \bf {v}\Vert _{TV}+{\frac{\rho }{2}\Vert {\bf {v}-\bf {m}_{k+1}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(c) |$\bf {v}_{2}=\underset{\bf {v}}{\arg \min }\lbrace \beta \Phi (\bf {v})+{\frac{\rho }{2}\Vert {\bf {v}-{\bf {v}_{1}}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(d) |$\bf {v}_{3}=\underset{\bf {v}}{\arg \min }\lbrace \gamma \mathcal {D}(\bf {v})+{\frac{\rho }{2}\Vert {\bf {v}-{\bf {v}_{2}}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(e) Let |$\bf {v}^{k+1}=\bf {v}_{3}$|⁠, |$\bf {u}_{k+1}=\bf {u}_{k}+\rho (\bf {m}_{k+1}-\bf {v}^{k+1})$|⁠;
End For
Output result|$\bf {m}_{K}$|⁠.
Initialization:Pick initial guess |$\bf {u}_{0}$|⁠, |$\bf {v}^{0}$|⁠, parameters |$\rho$|⁠, |$\alpha$|⁠, |$\beta$|⁠, and |$\gamma$|⁠.
For |$k=0,1,\cdots ,K-1$|
(a) |$\bf {m}_{k+1}=\underset{\bf {m}}{\arg \min }\lbrace \frac{1}{2}\Vert \mathcal {F}(\bf {m})-\bf {d}\Vert _{2}^{2}+ {\frac{\rho }{2}\Vert {\bf {m}-\bf {v}^{k}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace ;$|
(b) |$\bf {v}_{1}=\underset{\bf {v}}{\arg \min }\lbrace \alpha \Vert \bf {v}\Vert _{TV}+{\frac{\rho }{2}\Vert {\bf {v}-\bf {m}_{k+1}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(c) |$\bf {v}_{2}=\underset{\bf {v}}{\arg \min }\lbrace \beta \Phi (\bf {v})+{\frac{\rho }{2}\Vert {\bf {v}-{\bf {v}_{1}}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(d) |$\bf {v}_{3}=\underset{\bf {v}}{\arg \min }\lbrace \gamma \mathcal {D}(\bf {v})+{\frac{\rho }{2}\Vert {\bf {v}-{\bf {v}_{2}}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(e) Let |$\bf {v}^{k+1}=\bf {v}_{3}$|⁠, |$\bf {u}_{k+1}=\bf {u}_{k}+\rho (\bf {m}_{k+1}-\bf {v}^{k+1})$|⁠;
End For
Output result|$\bf {m}_{K}$|⁠.
Algorithm 1

PPR method

Initialization:Pick initial guess |$\bf {u}_{0}$|⁠, |$\bf {v}^{0}$|⁠, parameters |$\rho$|⁠, |$\alpha$|⁠, |$\beta$|⁠, and |$\gamma$|⁠.
For |$k=0,1,\cdots ,K-1$|
(a) |$\bf {m}_{k+1}=\underset{\bf {m}}{\arg \min }\lbrace \frac{1}{2}\Vert \mathcal {F}(\bf {m})-\bf {d}\Vert _{2}^{2}+ {\frac{\rho }{2}\Vert {\bf {m}-\bf {v}^{k}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace ;$|
(b) |$\bf {v}_{1}=\underset{\bf {v}}{\arg \min }\lbrace \alpha \Vert \bf {v}\Vert _{TV}+{\frac{\rho }{2}\Vert {\bf {v}-\bf {m}_{k+1}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(c) |$\bf {v}_{2}=\underset{\bf {v}}{\arg \min }\lbrace \beta \Phi (\bf {v})+{\frac{\rho }{2}\Vert {\bf {v}-{\bf {v}_{1}}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(d) |$\bf {v}_{3}=\underset{\bf {v}}{\arg \min }\lbrace \gamma \mathcal {D}(\bf {v})+{\frac{\rho }{2}\Vert {\bf {v}-{\bf {v}_{2}}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(e) Let |$\bf {v}^{k+1}=\bf {v}_{3}$|⁠, |$\bf {u}_{k+1}=\bf {u}_{k}+\rho (\bf {m}_{k+1}-\bf {v}^{k+1})$|⁠;
End For
Output result|$\bf {m}_{K}$|⁠.
Initialization:Pick initial guess |$\bf {u}_{0}$|⁠, |$\bf {v}^{0}$|⁠, parameters |$\rho$|⁠, |$\alpha$|⁠, |$\beta$|⁠, and |$\gamma$|⁠.
For |$k=0,1,\cdots ,K-1$|
(a) |$\bf {m}_{k+1}=\underset{\bf {m}}{\arg \min }\lbrace \frac{1}{2}\Vert \mathcal {F}(\bf {m})-\bf {d}\Vert _{2}^{2}+ {\frac{\rho }{2}\Vert {\bf {m}-\bf {v}^{k}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace ;$|
(b) |$\bf {v}_{1}=\underset{\bf {v}}{\arg \min }\lbrace \alpha \Vert \bf {v}\Vert _{TV}+{\frac{\rho }{2}\Vert {\bf {v}-\bf {m}_{k+1}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(c) |$\bf {v}_{2}=\underset{\bf {v}}{\arg \min }\lbrace \beta \Phi (\bf {v})+{\frac{\rho }{2}\Vert {\bf {v}-{\bf {v}_{1}}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(d) |$\bf {v}_{3}=\underset{\bf {v}}{\arg \min }\lbrace \gamma \mathcal {D}(\bf {v})+{\frac{\rho }{2}\Vert {\bf {v}-{\bf {v}_{2}}+\bf {u}_{k}}\Vert _{2}^{2}}\rbrace$|⁠;
(e) Let |$\bf {v}^{k+1}=\bf {v}_{3}$|⁠, |$\bf {u}_{k+1}=\bf {u}_{k}+\rho (\bf {m}_{k+1}-\bf {v}^{k+1})$|⁠;
End For
Output result|$\bf {m}_{K}$|⁠.

In fact, in the PPR method, each regularization subproblem corresponds to a denoising problem (Brifman et al.2016). Generally speaking, the proximal map

is equivalent to a denoising operation, where |$\bf {m}$| is a noisy intermediate solution, and |$\mathcal {R}$| denotes a regularizer corresponding to a denoiser with a denoising threshold |$\sigma$|⁠. In this paper, the denoising thresholds of denoising steps (b)–(d) are represented by |$\sigma _{i}(i=1,2,3)$| respectively, which are decided by the regularization parameter and penalty parameter |$\rho$| through the following equations:

In particular, in the implementation of the above denoising operations, it can be observed that the current denoising results are integrated into the noisy image of the next denoising step in Algorithm 1 to improve the inversion efficiency. Therefore, inspired by the spirit of the proposed PPR method, the solutions of optimization problems (b)–(d) of denoising stage can be denoted by

3. Applications in full waveform inversion

3.1. Acoustic full waveform inversion

Mathematically, FWI is a typical nonlinear ill-posed problem, which seeks to obtain a physical model that agrees with observed seismic data as well as possible.

Succinctly, frequency-domain FWI can be written as the following nonlinear least-squares optimization problem

(5)

where |$\bf {m}$| denotes the velocity model to be inverted, |$\omega$| is the angular frequency, |$A_{\omega }(\bf {m})=({\omega ^{2}}/{\bf {m}^{2}}) +\nabla ^{2}$| is the Helmholtz operator with a perfectly matched layer (PML) boundary condition, |$\bf {d}$| is the observed seismic data, |$\bf {Q}$| is the source matrix, and |$\bf {P}$| denotes a measurement operator that projects the solution of the Helmholtz equation on to the receiver locations. In this paper, we ignore the density parameter and only invert for acoustic velocity. In the discrete case, the unknown velocity model |$\bf {m}$| is a matrix of dimension |$N_{z}\times N_{x}$|⁠, where |$N_{z}$| and |$N_{x}$| are the numbers of grid points along the z and x directions, respectively. The limited memory quasi-Newton L-BFGS method (Zhu et al.1997) is utilized to solve step (a) in Algorithm 1 when the gradient of |$\mathcal {F}(\bf {m})={\bf {P}}A_{\omega }(\bf {m})^{-1}\bf {Q}$| is computed via the adjoint-state method (Plessix 2006).

The developed PPR algorithm for solving the acoustic FWI problem is illustrated in the flowchart in figure 1. Algorithm 2 (called the PPR–FWI algorithm) summarizes the PPR regularization procedure that allows a computer to solve FWI problems (5).

The flow chart of the PPR-FWI algorithm.
Figure 1.

The flow chart of the PPR-FWI algorithm.

Algorithm 2

PPR–FWI method

Initialize:
Given initial guess |$\bf {m}_{0}$|⁠, the observed data |$\bf {d}$|⁠, the maximum
number of iterations L, |$\bf {v}^{0}=\bf {u}_{0}=\rho _{0}=0$|⁠,
parameters |$\alpha$|⁠, |$\beta$|⁠, |$\gamma$|⁠, and |$\epsilon$|⁠.
Iterate: for |$l=0,1, \cdots ,L-1$|
(1) L-BFGS step:
|$\bf {m}_{l+1}=\underset{\bf {m}}{\arg \min }\lbrace {\frac{1}{2}}\Vert {\bf {P}}A_{\omega }(\bf {m})^{-1}\bf {Q}-\bf {d}\Vert _{2}^{2}+{\frac{\rho _{l}}{2}\Vert {\bf {m}-\bf {v}^{l}+\bf {u}_{l}}\Vert _{2}^{2}}\rbrace$|⁠;
(2) Denoising steps:
 (a) |$\bf {v}_{1}=\mbox{TV}(\bf {m}_{l+1}-\bf {u}_{l},\sqrt{\frac{\alpha }{\rho _{l}}})$|⁠;
 (b) |$\bf {v}_{2}=\mbox{BM3D}(\bf {v}_{1}-\bf {u}_{l},\sqrt{\frac{\beta }{\rho _{l}}})$|⁠;
 (c) |$\bf {v}_{3}=\mbox{FFDNet}(\bf {v}_{2}-\bf {u}_{l},\sqrt{\frac{\gamma }{\rho _{l}}})$|⁠.
(3) Update step:
 Let |$\bf {v}^{l+1}=\bf {v}_{3}$|⁠,
|$\rho _{l+1}=(l+1)(1+\epsilon )^{l+1}$|⁠,
|$\bf {u}_{l+1}=\bf {u}_{l}+\rho _{l+1}(\bf {m}_{l+1}-\bf {v}^{l+1})$|⁠;
Output: The inversion result is |${\bf {v}}^{L}$|⁠.
Initialize:
Given initial guess |$\bf {m}_{0}$|⁠, the observed data |$\bf {d}$|⁠, the maximum
number of iterations L, |$\bf {v}^{0}=\bf {u}_{0}=\rho _{0}=0$|⁠,
parameters |$\alpha$|⁠, |$\beta$|⁠, |$\gamma$|⁠, and |$\epsilon$|⁠.
Iterate: for |$l=0,1, \cdots ,L-1$|
(1) L-BFGS step:
|$\bf {m}_{l+1}=\underset{\bf {m}}{\arg \min }\lbrace {\frac{1}{2}}\Vert {\bf {P}}A_{\omega }(\bf {m})^{-1}\bf {Q}-\bf {d}\Vert _{2}^{2}+{\frac{\rho _{l}}{2}\Vert {\bf {m}-\bf {v}^{l}+\bf {u}_{l}}\Vert _{2}^{2}}\rbrace$|⁠;
(2) Denoising steps:
 (a) |$\bf {v}_{1}=\mbox{TV}(\bf {m}_{l+1}-\bf {u}_{l},\sqrt{\frac{\alpha }{\rho _{l}}})$|⁠;
 (b) |$\bf {v}_{2}=\mbox{BM3D}(\bf {v}_{1}-\bf {u}_{l},\sqrt{\frac{\beta }{\rho _{l}}})$|⁠;
 (c) |$\bf {v}_{3}=\mbox{FFDNet}(\bf {v}_{2}-\bf {u}_{l},\sqrt{\frac{\gamma }{\rho _{l}}})$|⁠.
(3) Update step:
 Let |$\bf {v}^{l+1}=\bf {v}_{3}$|⁠,
|$\rho _{l+1}=(l+1)(1+\epsilon )^{l+1}$|⁠,
|$\bf {u}_{l+1}=\bf {u}_{l}+\rho _{l+1}(\bf {m}_{l+1}-\bf {v}^{l+1})$|⁠;
Output: The inversion result is |${\bf {v}}^{L}$|⁠.
Algorithm 2

PPR–FWI method

Initialize:
Given initial guess |$\bf {m}_{0}$|⁠, the observed data |$\bf {d}$|⁠, the maximum
number of iterations L, |$\bf {v}^{0}=\bf {u}_{0}=\rho _{0}=0$|⁠,
parameters |$\alpha$|⁠, |$\beta$|⁠, |$\gamma$|⁠, and |$\epsilon$|⁠.
Iterate: for |$l=0,1, \cdots ,L-1$|
(1) L-BFGS step:
|$\bf {m}_{l+1}=\underset{\bf {m}}{\arg \min }\lbrace {\frac{1}{2}}\Vert {\bf {P}}A_{\omega }(\bf {m})^{-1}\bf {Q}-\bf {d}\Vert _{2}^{2}+{\frac{\rho _{l}}{2}\Vert {\bf {m}-\bf {v}^{l}+\bf {u}_{l}}\Vert _{2}^{2}}\rbrace$|⁠;
(2) Denoising steps:
 (a) |$\bf {v}_{1}=\mbox{TV}(\bf {m}_{l+1}-\bf {u}_{l},\sqrt{\frac{\alpha }{\rho _{l}}})$|⁠;
 (b) |$\bf {v}_{2}=\mbox{BM3D}(\bf {v}_{1}-\bf {u}_{l},\sqrt{\frac{\beta }{\rho _{l}}})$|⁠;
 (c) |$\bf {v}_{3}=\mbox{FFDNet}(\bf {v}_{2}-\bf {u}_{l},\sqrt{\frac{\gamma }{\rho _{l}}})$|⁠.
(3) Update step:
 Let |$\bf {v}^{l+1}=\bf {v}_{3}$|⁠,
|$\rho _{l+1}=(l+1)(1+\epsilon )^{l+1}$|⁠,
|$\bf {u}_{l+1}=\bf {u}_{l}+\rho _{l+1}(\bf {m}_{l+1}-\bf {v}^{l+1})$|⁠;
Output: The inversion result is |${\bf {v}}^{L}$|⁠.
Initialize:
Given initial guess |$\bf {m}_{0}$|⁠, the observed data |$\bf {d}$|⁠, the maximum
number of iterations L, |$\bf {v}^{0}=\bf {u}_{0}=\rho _{0}=0$|⁠,
parameters |$\alpha$|⁠, |$\beta$|⁠, |$\gamma$|⁠, and |$\epsilon$|⁠.
Iterate: for |$l=0,1, \cdots ,L-1$|
(1) L-BFGS step:
|$\bf {m}_{l+1}=\underset{\bf {m}}{\arg \min }\lbrace {\frac{1}{2}}\Vert {\bf {P}}A_{\omega }(\bf {m})^{-1}\bf {Q}-\bf {d}\Vert _{2}^{2}+{\frac{\rho _{l}}{2}\Vert {\bf {m}-\bf {v}^{l}+\bf {u}_{l}}\Vert _{2}^{2}}\rbrace$|⁠;
(2) Denoising steps:
 (a) |$\bf {v}_{1}=\mbox{TV}(\bf {m}_{l+1}-\bf {u}_{l},\sqrt{\frac{\alpha }{\rho _{l}}})$|⁠;
 (b) |$\bf {v}_{2}=\mbox{BM3D}(\bf {v}_{1}-\bf {u}_{l},\sqrt{\frac{\beta }{\rho _{l}}})$|⁠;
 (c) |$\bf {v}_{3}=\mbox{FFDNet}(\bf {v}_{2}-\bf {u}_{l},\sqrt{\frac{\gamma }{\rho _{l}}})$|⁠.
(3) Update step:
 Let |$\bf {v}^{l+1}=\bf {v}_{3}$|⁠,
|$\rho _{l+1}=(l+1)(1+\epsilon )^{l+1}$|⁠,
|$\bf {u}_{l+1}=\bf {u}_{l}+\rho _{l+1}(\bf {m}_{l+1}-\bf {v}^{l+1})$|⁠;
Output: The inversion result is |${\bf {v}}^{L}$|⁠.

3.2. Numerical experiments

The quality of the inverted velocity models is measured using the peak signal-to-noise ratio (PSNR), the structural similarity index measure (SSIM), and the root mean square error (RMSE). We compared the proposed PPR–FWI method with three other algorithms: (1) FWI with the TV denoiser (TV–FWI); (2) FWI with the BM3D denoiser (BM3D–FWI); (3) FWI with FFDNet networks (FFDNet–FWI). In the experiments, we set the outer loop as |$L=4$| and employ 10 iterations in the L-BFGS iteration stage, which is consistent among four FWI methods to ensure a fair and valid comparison. From a practical perspective, noise is inevitable in the collected seismic data. In all experiments, we also consider a noise version of the seismic data |$\bf {d}$|⁠, to which 5% white Gaussian noise is added. In the frequency domain FWI, we parallelize over frequencies and invert all frequency data components simultaneously. In the simulation, the FFDNet model and parameters from https://github.com/cszn/FFDNet are directly used.

In the PPR–FWI method, to satisfy the condition required in the convergence theorem stated in Cascarano et al. (2022), we choose |$\rho _{l+1}=(l+1)(1+\epsilon )^{l+1}$| with |$\epsilon \gt 0$|⁠; pertinent details can be found in the appendix section of Cascarano et al. (2022). For the parameter |$\epsilon$|⁠, we find that |$\epsilon =10^{-3}$| will lead to a good numerical performance. The three regularization parameters for the PPR–FWI algorithm are set as |$\alpha =1\times 10^{-3}$|⁠, |$\beta =0.02$|⁠, and |$\gamma =10^{-4}$|⁠, respectively. In addition, the regularization parameters of comparative inversion methods for different numerical examples can vary, which will be described in the following subsections.

3.2.1. Marmousi model

We perform the modified Marmousi synthetic benchmark model as the first numerical example. The dimensions of the model are |$3060 \times 9600$| m|$^2$|⁠, with grid spacings of 4 m, as shown in figure 2(a). The frequency bandwidth is [2.9,24] Hz. The initial velocity model (figure 2(b)) is estimated by smoothing the original model. We place |$N_{s}=75$| sources and |$N_{r}=225$| receivers located uniformly on the model surface. For the implementation of the TV–FWI, BM3D–FWI, and FFDNet–FWI methods, the regularization parameters are set to 0.8, 0.2, and 0.15, respectively.

The true velocity and initial velocity of the Marmousi model.
Figure 2.

The true velocity and initial velocity of the Marmousi model.

The inversion results with four different regularization schemes are shown in figures 3 and 4. Specifically, figure 3 shows inversion results without Gaussian noise in the data, and results with noise are given in figure 4. As shown in figures 3 and 4, the inversion results of TV–FWI, BM3D–FWI, and FFDNet–FWI produce a reasonable velocity model, while the features in the deeper part of the model are poorly restored. By contrast, FWI with PPR regularization improves the reconstruction quality for seismic imaging to some extent. The inverted model obtained by the PPR–FWI method (depicted in figures 3(d) and 4(d)) contains more structural details and fewer inversion artifacts than other methods, which further demonstrates that our PPR–FWI algorithm obtains the best performance. A comparison of the velocity profiles shown in figures 5 and 6 further demonstrates the superiority of our PPR–FWI method.

The inversion results without noise in seismic data (Marmousi model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR–FWI.
Figure 3.

The inversion results without noise in seismic data (Marmousi model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR–FWI.

The inversion results with 5% white noise in seismic data (Marmousi model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR–FWI.
Figure 4.

The inversion results with 5% white noise in seismic data (Marmousi model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR–FWI.

The velocity profiles extracted from the Marmousi model without noise in seismic data.
Figure 5.

The velocity profiles extracted from the Marmousi model without noise in seismic data.

The velocity profiles extracted from the Marmousi model with noise in seismic data.
Figure 6.

The velocity profiles extracted from the Marmousi model with noise in seismic data.

To further illustrate the performance of the proposed algorithm, the quantitative results of the Marmousi model are reported in Table 1. We can observe that the proposed algorithm realizes a relatively highly competitive performance. It basically outperforms the other three algorithms according to PSNR, SSIM, and RMSE values.

Table 1.

Quantitative evaluation by different algorithms for the Marmousi model.

Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize19.190.620.1119.190.620.11
TV–FWI23.250.780.0723.200.780.07
BM3D–FWI23.570.790.0723.510.780.07
FFDNet–FWI22.860.750.0722.790.750.07
PPR–FWI23.810.800.0623.750.800.06
Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize19.190.620.1119.190.620.11
TV–FWI23.250.780.0723.200.780.07
BM3D–FWI23.570.790.0723.510.780.07
FFDNet–FWI22.860.750.0722.790.750.07
PPR–FWI23.810.800.0623.750.800.06
Table 1.

Quantitative evaluation by different algorithms for the Marmousi model.

Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize19.190.620.1119.190.620.11
TV–FWI23.250.780.0723.200.780.07
BM3D–FWI23.570.790.0723.510.780.07
FFDNet–FWI22.860.750.0722.790.750.07
PPR–FWI23.810.800.0623.750.800.06
Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize19.190.620.1119.190.620.11
TV–FWI23.250.780.0723.200.780.07
BM3D–FWI23.570.790.0723.510.780.07
FFDNet–FWI22.860.750.0722.790.750.07
PPR–FWI23.810.800.0623.750.800.06

3.2.2. Modified BP 2004 model

As a second example, a numerical experiment based on a modified BP 2004 Model (see figure 7(a)) is presented to test the performance of the PPR method. The model is discretized to |$N_{z}\times N_{x}=128\times 512$| with a grid spacing of 8 m. The frequency bandwidth is [2,26] Hz. The initial model (see figure 7(b)) is obtained by applying Gaussian smoothing to the true model. We evenly distribute 43 sources and 129 receivers in the horizontal direction. The manually selected regularization parameters for the TV–FWI, BM3D–FWI, and FFDNet–FWI methods are set as |$10^{-4}$|⁠, 0.5, and |$6\times 10^{-3}$|⁠, respectively.

The true velocity and initial velocity of the BP 2004 model.
Figure 7.

The true velocity and initial velocity of the BP 2004 model.

Figures 8 and 9 show the inversion velocity model with four different regularization methods under noiseless and noise conditions. If we compare the inversion results carefully, in the areas shown by the rectangular boxes, the PPR–FWI method (depicted in figures 8(d) and 9(d)) can better reconstruct the detailed features and reveal the “tooth” part of the velocity model. Figures 10 and 11 show the velocity profiles of the inversion results at |$z=960$| m under noise-free and noisy cases, respectively. Compared with the TV–FWI, BM3D–FWI, and FFDNet–FWI methods, the inversion result obtained by the proposed PPR–FWI algorithm is closer to the true velocity of the model. Furthermore, Table 2 shows the quantitative inversion results. As can be seen from these metrics, the proposed PPR–FWI method achieves a satisfactory improvement, which further indicates that prior information utilized in the PPR–FWI algorithm can obtain the highest reconstruction quality.

The inversion results without noise in seismic data (BP 2004 model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR-FWI.
Figure 8.

The inversion results without noise in seismic data (BP 2004 model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR-FWI.

The inversion results with 5% white noise in seismic data (BP 2004 model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR–FWI.
Figure 9.

The inversion results with 5% white noise in seismic data (BP 2004 model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR–FWI.

The velocity profiles extracted from the BP 2004 model without noise in seismic data.
Figure 10.

The velocity profiles extracted from the BP 2004 model without noise in seismic data.

The velocity profiles extracted from the BP 2004 model with noise in seismic data.
Figure 11.

The velocity profiles extracted from the BP 2004 model with noise in seismic data.

Table 2.

Quantitative evaluation by different algorithms for the modified BP 2004 model.

Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize10.980.710.2810.980.710.28
TV–FWI21.080.820.0918.260.690.12
BM3D–FWI19.040.870.1018.700.840.12
FFDNet–FWI20.530.880.0918.670.820.12
PPR–FWI21.990.910.0820.180.870.10
Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize10.980.710.2810.980.710.28
TV–FWI21.080.820.0918.260.690.12
BM3D–FWI19.040.870.1018.700.840.12
FFDNet–FWI20.530.880.0918.670.820.12
PPR–FWI21.990.910.0820.180.870.10
Table 2.

Quantitative evaluation by different algorithms for the modified BP 2004 model.

Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize10.980.710.2810.980.710.28
TV–FWI21.080.820.0918.260.690.12
BM3D–FWI19.040.870.1018.700.840.12
FFDNet–FWI20.530.880.0918.670.820.12
PPR–FWI21.990.910.0820.180.870.10
Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize10.980.710.2810.980.710.28
TV–FWI21.080.820.0918.260.690.12
BM3D–FWI19.040.870.1018.700.840.12
FFDNet–FWI20.530.880.0918.670.820.12
PPR–FWI21.990.910.0820.180.870.10

3.2.3. SEG/EAGE salt model

We finally experiment by using our method on a two-dimensional SEG/EAGE salt model. The true model is depicted in figure 12(a), which is discretized to |$N_{z}\times N_{x}=128\times 256$| with a spacing of 5 m. The frequency bandwidth is [2,26] Hz. The synthetic data are generated by 31 sources and 93 receivers near the surface. Figure 12(b) shows the initial velocity model. Note that, to mitigate the reliance on the initial value, there is a total lack of information about lateral variation in the initial velocity model. For the implementation of TV–FWI, BM3D–FWI, and FFDNet–FWI methods, the corresponding regularization parameters are set as |$2\times 10^{-4}$|⁠, 0.5, and |$5\times 10^{-5}$|⁠, respectively.

The true velocity and initial velocity of the SEG/EAGE salt model.
Figure 12.

The true velocity and initial velocity of the SEG/EAGE salt model.

Similarly, this experiment is also studied in the noise-free and noisy cases (see figures 13 and 14). We make the following conclusions in terms of visual effects and performance metrics. In the noiseless situation, as shown in figures 13(a) and (c), the inversion of TV–FWI and FFDNet–FWI leads to lower visual clarity while strong artifacts occur in the background part. By examining the inversion result obtained by the BM3D–FWI method (figure 13(b)), the salt structure is better maintained while the background part of the inverted model is poor. The inversion results for the PPR–FWI method are shown in figure 13(d). Notably, we find that our proposed method results in a more accurate reconstructed image, and the structural details and boundaries of the model are well preserved. In the noisy case, the inversion results obtained by the proposed PPR–FWI algorithm (see figure 14(d)) generate the slightest deviation with a more precise background and more obvious observation of the inverted model.

The inversion results without noise in seismic data (SEG/EAGE salt model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR–FWI.
Figure 13.

The inversion results without noise in seismic data (SEG/EAGE salt model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR–FWI.

The inversion results with 5% white noise in seismic data (SEG/EAGE salt model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR–FWI.
Figure 14.

The inversion results with 5% white noise in seismic data (SEG/EAGE salt model): (a) TV–FWI; (b) BM3D–FWI; (c) FFDNet–FWI; (d) PPR–FWI.

Figures 15 and 16 show the velocity profiles of the inversion results obtained by four inversion algorithms at |$x=900$| m. Although the inversion velocities of the four algorithms are all close to the true velocity, the PPR–FWI algorithm obtains better fitting effects between the inversion velocity and the true velocity. Table 3 shows the quantitative evaluation of the reconstructed results. As can be seen from Table 3, our PPR–FWI method achieves the best performance for the quantitative metrics measured by PSNR, SSIM, and RMSE.

The velocity profiles extracted from the SEG/EAGE salt model without noise in seismic data.
Figure 15.

The velocity profiles extracted from the SEG/EAGE salt model without noise in seismic data.

The velocity profiles extracted from the SEG/EAGE salt model with noise in seismic data.
Figure 16.

The velocity profiles extracted from the SEG/EAGE salt model with noise in seismic data.

Table 3.

Quantitative evaluation by different algorithms for the SEG/EAGE salt model.

Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize5.860.430.515.860.430.51
TV–FWI11.840.580.2611.310.540.27
BM3D–FWI11.770.550.2611.340.530.27
FFDNet–FWI11.760.580.2611.270.530.27
PPR–FWI12.030.620.2511.480.580.26
Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize5.860.430.515.860.430.51
TV–FWI11.840.580.2611.310.540.27
BM3D–FWI11.770.550.2611.340.530.27
FFDNet–FWI11.760.580.2611.270.530.27
PPR–FWI12.030.620.2511.480.580.26
Table 3.

Quantitative evaluation by different algorithms for the SEG/EAGE salt model.

Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize5.860.430.515.860.430.51
TV–FWI11.840.580.2611.310.540.27
BM3D–FWI11.770.550.2611.340.530.27
FFDNet–FWI11.760.580.2611.270.530.27
PPR–FWI12.030.620.2511.480.580.26
Noise-free caseNoisy case
PSNRSSIMRMSEPSNRSSIMRMSE
Initialize5.860.430.515.860.430.51
TV–FWI11.840.580.2611.310.540.27
BM3D–FWI11.770.550.2611.340.530.27
FFDNet–FWI11.760.580.2611.270.530.27
PPR–FWI12.030.620.2511.480.580.26

4. Conclusions

We design a flexible PnP regularization algorithm by incorporating the sparsity and nonlocal self-similarity priors, as well as the deep priors learned by deep denoiser networks. What is more, three powerful denoisers (TV, BM3D, and FFDNet) are plugged to solve regularization subproblems. The application of the PnP regularization algorithm in the frequency-domain FWI problem leads to promising performance. In the experiments based on the Marmousi model, SEG/EAGE salt model, and modified BP 2004 model, our algorithm shows a superior ability in reducing artifacts and preserving more details to a great extent. Furthermore, because performing L-BFGS iterations in the FWI process is the most time-consuming stage, the computation time for all inversion methods is almost the same. However, the proposed method still has room for improvement in several aspects. The convergence analysis has drawn more attention and it is challenging to verify whether general denoisers meet the required conditions. In addition, the adaptive selection of regularization parameters and the improvement of computational efficiency still need further exploration in practical applications.

Acknowledgements

This work was supported by the National Natural Science Foundation of China (grant no. 42274166)

Data Availability

Data will be made available on request.

Conflict of interest statement. None.

References

Boyd
S.
,
Parikh
N.
,
Chu
E.
,
Peleato
B.
,
Eckstein
J.
,
2010
.
Distributed optimization and statistical learning via the alternating direction method of multipliers
,
Foundations & Trends in Machine Learning
,
3
,
1
122
.

Brifman
A.
,
Romano
Y.
,
Elad
M.
,
2016
.
Turning a denoiser into a super-resolver using plug and play priors
,
IEEE Int. Conf. Image Processing
,
3
,
1404
1408
.

Burger
H.C.
,
Schuler
C.J.
,
Harmeling
S.
,
2012
.
Image denoising: can plain neural networks compete with BM3D
,
2012 IEEE Conference on Computer Vision and Pattern Recognition
,
3
,
2392
2399
.

Cascarano
P.
,
Piccolomini
E.L.
,
Morotti
E.
,
Sebastiani
A.
,
2022
.
Plug-and-play gradient-based denoisers applied to CT image enhancement
,
Appl. Math. Comput.
,
422
,
126967
.

Chan
S.H.
,
Wang
X.
,
Elgendy
O.A.
,
2017
.
Plug-and-play ADMM for image restoration: fixed point convergence and applications
,
IEEE Trans. Comput. Imaging
,
3
,
84
98
.

Choi
Y.
,
Shin
C.
,
Min
D.-J.
,
Ha
T.
,
2005
.
Efficient calculation of the steepest descent direction for source-independent seismic waveform inversion: an amplitude approach
,
J. Comput. Phys.
,
208
,
455
468
.

Dabov
K.
,
Foi
A.
,
Katkovnik
V.
,
Egiazarian
K.O.
,
2007
.
Image denoising by sparse 3-D transform-domain collaborative filtering
,
IEEE Trans. Image Processing
,
16
,
2080
2095
.

Esser
E.
,
Guasch
L.
,
Leeuwen
T.V.
,
Aravkin
A.Y.
,
Herrmann
F.J.
,
2016
.
Total variation regularization strategies in full-waveform inversion
,
SIAM J. Imaging Sciences
,
11
,
376
406
.

Li
L.
,
Han
B.
,
2016
.
A new iteratively total variational regularization for nonlinear inverse problems
,
Appl. Math. Comput.
,
298
,
40
52
.

Lin
Y.
,
Huang
L.
,
2014
.
Acoustic- and elastic-waveform inversion using a modified total-variation regularization scheme
,
Geophys. J. Int.
,
200
,
489
502
.

Peng
Y.
,
Liao
W.
,
Huang
J.
,
Li
Z.
,
2018
.
Total variation regularization for seismic waveform inversion using an adaptive primal dual hybrid gradient method
,
Inverse Problems
,
34
,
045006
.

Plessix
R.E.
,
2006
.
A review of the adjoint-state method for computing the gradient of a functional with geophysical applications
,
Geophys. J. Int.
,
167
,
495
503
.

Rasti-Meymandi
A.
,
Ghaffari
A.
,
Fatemizadeh
E.
,
2022
.
Plug and play augmented HQS: convergence analysis and its application in MRI reconstruction
,
Neurocomputing
,
518
,
1
14
.

Rudin
L.I.
,
Osher
S.
,
Fatemi
E.
,
1992
.
Nonlinear total variation based noise removal algorithms
,
Physica D Nonlinear Phenomena
,
60
,
259
268
.

Sun
J.
,
Niu
Z.
,
Innanen
K.A.
,
Li
J.
,
Trad
D.O.
,
2019
.
A theory-guided deep learning formulation and optimization of seismic waveform inversion
,
Geophysics
,
85
,
1
63
.

Tikhonov
A.N.
,
Arsenin
V.Y.
,
1977
.
Solutions of ill-posed problems
,
Math. Comput.
,
32
,
491
491
.

Venkatakrishnan
S.
,
Bouman
C.A.
,
Wohlberg
B.
,
2013
.
Plug-and-play priors for model based reconstruction
,
IEEE Global Conference on Signal and Information Processing
,
3
,
945
948
.

Wang
Y.
,
Seismic Inversion, Theory and Applications
,
Oxford: Wiley Blackwell
,
2016
.

Wu
Y.
,
Lin
Y.
,
2020
.
Inversionnet: an efficient and accurate data-driven full waveform inversion
,
IEEE Transactions on Computational Imaging
,
6
,
419
433
.

Zhang
K.
,
Zuo
W.
,
Zhang
L.
,
2018
.
FFDNet: toward a fast and flexible solution for CNN-based image denoising
,
IEEE Trans. Image Processing
,
27
,
4608
4622
.

Zhang
Z.D.
,
Alkhalifah
T.
,
2019
.
Regularized elastic full waveform inversion using deep learning
,
Geophysics
,
84
,
1
47
.

Zhu
C.
,
Byrd
R.H.
,
Lu
P.
,
Nocedal
J.
,
1997
.
Algorithm 778: L-BFGS-B: Fortran subroutines for large-scale bound-constrained optimization
,
ACM Trans. Math. Software
,
23
,
550
560
.

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.