Sparse Inversion with Iteratively Re-Weighted Least-Squares#

Least-squares inversion produces smooth models which may not be an accurate representation of the true model. Here we demonstrate the basics of inverting for sparse and/or blocky models. Here, we used the iteratively reweighted least-squares approach. For this tutorial, we focus on the following:

  • Defining the forward problem

  • Defining the inverse problem (data misfit, regularization, optimization)

  • Defining the paramters for the IRLS algorithm

  • Specifying directives for the inversion

  • Recovering a set of model parameters which explains the observations

import numpy as np
import matplotlib.pyplot as plt

from discretize import TensorMesh

from simpeg import (
    simulation,
    maps,
    data_misfit,
    directives,
    optimization,
    regularization,
    inverse_problem,
    inversion,
)

# sphinx_gallery_thumbnail_number = 3

Defining the Model and Mapping#

Here we generate a synthetic model and a mappig which goes from the model space to the row space of our linear operator.

nParam = 100  # Number of model paramters

# A 1D mesh is used to define the row-space of the linear operator.
mesh = TensorMesh([nParam])

# Creating the true model
true_model = np.zeros(mesh.nC)
true_model[mesh.cell_centers_x > 0.3] = 1.0
true_model[mesh.cell_centers_x > 0.45] = -0.5
true_model[mesh.cell_centers_x > 0.6] = 0

# Mapping from the model space to the row space of the linear operator
model_map = maps.IdentityMap(mesh)

# Plotting the true model
fig = plt.figure(figsize=(8, 5))
ax = fig.add_subplot(111)
ax.plot(mesh.cell_centers_x, true_model, "b-")
ax.set_ylim([-2, 2])
plot inv 2 inversion irls
(-2.0, 2.0)

Defining the Linear Operator#

Here we define the linear operator with dimensions (nData, nParam). In practive, you may have a problem-specific linear operator which you would like to construct or load here.

# Number of data observations (rows)
nData = 20

# Create the linear operator for the tutorial. The columns of the linear operator
# represents a set of decaying and oscillating functions.
jk = np.linspace(1.0, 60.0, nData)
p = -0.25
q = 0.25


def g(k):
    return np.exp(p * jk[k] * mesh.cell_centers_x) * np.cos(
        np.pi * q * jk[k] * mesh.cell_centers_x
    )


G = np.empty((nData, nParam))

for i in range(nData):
    G[i, :] = g(i)

# Plot the columns of G
fig = plt.figure(figsize=(8, 5))
ax = fig.add_subplot(111)
for i in range(G.shape[0]):
    ax.plot(G[i, :])

ax.set_title("Columns of matrix G")
Columns of matrix G
Text(0.5, 1.0, 'Columns of matrix G')

Defining the Simulation#

The simulation defines the relationship between the model parameters and predicted data.

sim = simulation.LinearSimulation(mesh, G=G, model_map=model_map)

Predict Synthetic Data#

Here, we use the true model to create synthetic data which we will subsequently invert.

# Standard deviation of Gaussian noise being added
std = 0.02
np.random.seed(1)

# Create a SimPEG data object
data_obj = sim.make_synthetic_data(true_model, noise_floor=std, add_noise=True)

Define the Inverse Problem#

The inverse problem is defined by 3 things:

  1. Data Misfit: a measure of how well our recovered model explains the field data

  2. Regularization: constraints placed on the recovered model and a priori information

  3. Optimization: the numerical approach used to solve the inverse problem

# Define the data misfit. Here the data misfit is the L2 norm of the weighted
# residual between the observed data and the data predicted for a given model.
# Within the data misfit, the residual between predicted and observed data are
# normalized by the data's standard deviation.
dmis = data_misfit.L2DataMisfit(simulation=sim, data=data_obj)

# Define the regularization (model objective function). Here, 'p' defines the
# the norm of the smallness term and 'q' defines the norm of the smoothness
# term.
reg = regularization.Sparse(mesh, mapping=model_map)
reg.reference_model = np.zeros(nParam)
p = 0.0
q = 0.0
reg.norms = [p, q]

# Define how the optimization problem is solved.
opt = optimization.ProjectedGNCG(
    maxIter=100, lower=-2.0, upper=2.0, maxIterLS=20, cg_maxiter=30, cg_rtol=1e-3
)

# Here we define the inverse problem that is to be solved
inv_prob = inverse_problem.BaseInvProblem(dmis, reg, opt)

Define Inversion Directives#

Here we define any directiveas that are carried out during the inversion. This includes the cooling schedule for the trade-off parameter (beta), stopping criteria for the inversion and saving inversion results at each iteration.

# Add sensitivity weights but don't update at each beta
sensitivity_weights = directives.UpdateSensitivityWeights(every_iteration=False)

# Reach target misfit for L2 solution, then use IRLS until model stops changing.
IRLS = directives.UpdateIRLS(max_irls_iterations=40, f_min_change=1e-4)

# Defining a starting value for the trade-off parameter (beta) between the data
# misfit and the regularization.
starting_beta = directives.BetaEstimate_ByEig(beta0_ratio=1e0)

# Update the preconditionner
update_Jacobi = directives.UpdatePreconditioner()

# Save output at each iteration
saveDict = directives.SaveOutputEveryIteration(save_txt=False)

# Define the directives as a list
directives_list = [
    sensitivity_weights,
    IRLS,
    starting_beta,
    update_Jacobi,
    saveDict,
]
/home/vsts/work/1/s/simpeg/directives/_directives.py:1865: FutureWarning:

SaveEveryIteration.save_txt has been deprecated, please use SaveEveryIteration.on_disk. It will be removed in version 0.26.0 of SimPEG.

/home/vsts/work/1/s/simpeg/directives/_directives.py:1866: FutureWarning:

SaveEveryIteration.save_txt has been deprecated, please use SaveEveryIteration.on_disk. It will be removed in version 0.26.0 of SimPEG.

Setting a Starting Model and Running the Inversion#

To define the inversion object, we need to define the inversion problem and the set of directives. We can then run the inversion.

# Here we combine the inverse problem and the set of directives
inv = inversion.BaseInversion(inv_prob, directives_list)

# Starting model
starting_model = 1e-4 * np.ones(nParam)

# Run inversion
recovered_model = inv.run(starting_model)
Running inversion with SimPEG v0.25.2.dev2+gfcb9bdf36
================================================= Projected GNCG =================================================
  #     beta     phi_d     phi_m       f      |proj(x-g)-x|  LS   iter_CG   CG |Ax-b|/|b|  CG |Ax-b|   Comment
-----------------------------------------------------------------------------------------------------------------
   0  1.67e+06  3.76e+03  1.02e-09  3.76e+03                         0           inf          inf
   1  1.67e+06  1.91e+03  3.87e-04  2.56e+03    1.95e+01      0      8        3.41e-04     1.67e+00
   2  8.37e+05  1.32e+03  8.96e-04  2.07e+03    1.90e+01      0      9        2.73e-04     2.22e-01
   3  4.19e+05  7.81e+02  1.81e-03  1.54e+03    1.86e+01      0      9        9.21e-04     5.49e-01
   4  2.09e+05  3.96e+02  3.10e-03  1.04e+03    1.75e+01      0      11       6.36e-04     2.65e-01
   5  1.05e+05  1.78e+02  4.54e-03  6.54e+02    1.67e+01      0      13       7.28e-04     1.97e-01
   6  5.23e+04  7.67e+01  5.87e-03  3.84e+02    1.47e+01      0      17       6.87e-04     1.12e-01
   7  2.62e+04  3.63e+01  6.93e-03  2.18e+02    1.31e+01      0      21       9.39e-04     8.68e-02
   8  1.31e+04  2.12e+01  7.71e-03  1.22e+02    1.14e+01      0      30       7.39e-04     3.73e-02
   9  6.54e+03  1.53e+01  8.33e-03  6.98e+01    9.58e+00      0      30       3.94e-03     1.06e-01
Reached starting chifact with l2-norm regularization: Start IRLS steps...
irls_threshold 1.212904976635559
  10  6.54e+03  1.89e+01  1.04e-02  8.72e+01    1.36e+01      0      30       2.05e-03     3.43e-02
  11  6.54e+03  2.16e+01  1.14e-02  9.61e+01    9.90e+00      0      30       9.29e-04     9.43e-03
  12  6.54e+03  2.45e+01  1.20e-02  1.03e+02    1.00e+01      0      30       9.83e-04     1.03e-02
  13  5.12e+03  2.40e+01  1.29e-02  9.02e+01    3.85e+00      0      30       1.04e-03     6.81e-03
  14  4.06e+03  2.25e+01  1.36e-02  7.76e+01    4.81e+00      0      28       6.76e-04     4.57e-03
  15  3.32e+03  2.07e+01  1.36e-02  6.60e+01    5.09e+00      0      30       1.07e-03     7.24e-03
  16  3.32e+03  2.02e+01  1.26e-02  6.19e+01    5.76e+00      0      30       1.99e-03     1.24e-02
  17  3.32e+03  1.97e+01  1.14e-02  5.74e+01    6.20e+00      0      30       5.69e-04     3.78e-03
  18  3.32e+03  1.90e+01  1.01e-02  5.27e+01    6.62e+00      0      29       8.21e-04     5.90e-03
  19  3.32e+03  1.83e+01  8.92e-03  4.79e+01    6.78e+00      0      30       2.19e-04     1.63e-03
  20  3.32e+03  1.78e+01  7.91e-03  4.41e+01    7.57e+00      0      25       9.09e-04     7.63e-03
  21  5.18e+03  1.96e+01  6.34e-03  5.24e+01    1.39e+01      0      19       5.15e-04     1.50e-02
  22  5.18e+03  1.94e+01  5.47e-03  4.78e+01    8.10e+00      0      19       6.31e-04     6.63e-03
  23  5.18e+03  1.92e+01  4.79e-03  4.40e+01    8.72e+00      0      20       9.55e-04     1.13e-02
  24  5.18e+03  1.89e+01  4.26e-03  4.10e+01    9.42e+00      0      24       3.22e-04     4.40e-03
  25  5.18e+03  1.87e+01  3.80e-03  3.84e+01    9.66e+00      0      25       6.69e-04     9.66e-03
  26  5.18e+03  1.89e+01  3.37e-03  3.63e+01    9.84e+00      0      26       6.59e-04     9.75e-03
  27  5.18e+03  1.88e+01  2.96e-03  3.42e+01    1.67e+01      0      21       4.50e-04     1.43e-02
  28  5.18e+03  1.88e+01  2.53e-03  3.20e+01    8.55e+00      0      24       6.63e-04     8.73e-03
  29  5.18e+03  1.88e+01  2.17e-03  3.00e+01    8.56e+00      0      22       9.15e-04     1.22e-02
  30  5.18e+03  1.88e+01  1.86e-03  2.84e+01    8.63e+00      0      22       9.85e-04     1.62e-02
  31  5.18e+03  1.88e+01  1.60e-03  2.71e+01    9.08e+00      0      24       8.12e-04     1.19e-02
  32  5.18e+03  1.87e+01  1.37e-03  2.58e+01    1.53e+01      0      23       7.41e-04     1.57e-02
  33  5.18e+03  1.86e+01  1.17e-03  2.46e+01    9.81e+00      0      25       4.59e-04     7.39e-03
  34  5.18e+03  1.85e+01  9.83e-04  2.36e+01    1.01e+01      0      26       8.40e-04     1.40e-02
  35  5.18e+03  1.84e+01  8.22e-04  2.27e+01    1.01e+01      0      30       1.85e-03     3.17e-02
  36  5.18e+03  1.83e+01  6.84e-04  2.19e+01    1.01e+01      0      30       4.34e-03     7.66e-02
  37  5.18e+03  1.82e+01  5.70e-04  2.11e+01    9.78e+00      0      30       4.00e-03     7.23e-02
  38  5.18e+03  1.81e+01  4.68e-04  2.06e+01    9.63e+00      0      30       1.07e-02     2.04e-01
  39  5.18e+03  1.80e+01  3.79e-04  2.00e+01    1.59e+01      0      30       8.97e-03     2.41e-01
  40  8.06e+03  1.81e+01  2.95e-04  2.04e+01    1.34e+01      0      25       8.02e-04     6.74e-02
  41  8.06e+03  1.80e+01  2.49e-04  2.00e+01    1.06e+01      0      30       2.53e-02     7.78e-01
  42  1.25e+04  1.80e+01  1.97e-04  2.04e+01    1.33e+01      0      30       1.36e-03     1.38e-01
  43  1.95e+04  1.79e+01  1.60e-04  2.10e+01    1.33e+01      0      30       1.88e-03     2.21e-01
  44  3.04e+04  1.79e+01  1.31e-04  2.18e+01    1.36e+01      0      30       7.40e-03     9.45e-01
  45  4.75e+04  1.79e+01  1.07e-04  2.29e+01    1.42e+01      0      30       1.96e-02     2.65e+00
  46  7.41e+04  1.79e+01  8.81e-05  2.44e+01    1.48e+01      0      30       7.11e-03     1.00e+00
  47  1.16e+05  1.79e+01  7.21e-05  2.63e+01    1.54e+01      0      25       7.04e-04     1.00e-01
  48  1.80e+05  1.80e+01  5.93e-05  2.87e+01    1.62e+01      0      30       5.87e-03     1.05e+00
  49  1.80e+05  1.80e+01  5.02e-05  2.71e+01    1.11e+01      5      30       1.50e-01     4.32e+01
Reach maximum number of IRLS cycles: 40
------------------------- STOP! -------------------------
1 : |fc-fOld| = 7.4355e-04 <= tolF*(1+|f0|) = 3.7572e+02
1 : |xc-x_last| = 7.9978e-03 <= tolX*(1+|x0|) = 1.0010e-01
0 : |proj(x-g)-x|    = 1.1116e+01 <= tolG          = 1.0000e-01
0 : |proj(x-g)-x|    = 1.1116e+01 <= 1e3*eps       = 1.0000e-02
0 : maxIter   =     100    <= iter          =     49
------------------------- DONE! -------------------------

Plotting Results#

fig, ax = plt.subplots(1, 2, figsize=(12 * 1.2, 4 * 1.2))

# True versus recovered model
ax[0].plot(mesh.cell_centers_x, true_model, "k-")
ax[0].plot(mesh.cell_centers_x, inv_prob.l2model, "b-")
ax[0].plot(mesh.cell_centers_x, recovered_model, "r-")
ax[0].legend(("True Model", "Recovered L2 Model", "Recovered Sparse Model"))
ax[0].set_ylim([-2, 2])

# Observed versus predicted data
ax[1].plot(data_obj.dobs, "k-")
ax[1].plot(inv_prob.dpred, "ko")
ax[1].legend(("Observed Data", "Predicted Data"))

# Plot convergence
fig = plt.figure(figsize=(9, 5))
ax = fig.add_axes([0.2, 0.1, 0.7, 0.85])
ax.plot(saveDict.phi_d, "k", lw=2)

twin = ax.twinx()
twin.plot(saveDict.phi_m, "k--", lw=2)
ax.plot(
    np.r_[IRLS.metrics.start_irls_iter, IRLS.metrics.start_irls_iter],
    np.r_[0, np.max(saveDict.phi_d)],
    "k:",
)
ax.text(
    IRLS.metrics.start_irls_iter,
    0.0,
    "IRLS Start",
    va="bottom",
    ha="center",
    rotation="vertical",
    size=12,
    bbox={"facecolor": "white"},
)

ax.set_ylabel(r"$\phi_d$", size=16, rotation=0)
ax.set_xlabel("Iterations", size=14)
twin.set_ylabel(r"$\phi_m$", size=16, rotation=0)
  • plot inv 2 inversion irls
  • plot inv 2 inversion irls
Text(865.1527777777777, 0.5, '$\\phi_m$')

Total running time of the script: (0 minutes 25.740 seconds)

Estimated memory usage: 321 MB

Gallery generated by Sphinx-Gallery