Note
Click here to download the full example code
Sparse Inversion with Iteratively Re-Weighted Least-Squares#
Least-squares inversion produces smooth models which may not be an accurate representation of the true model. Here we demonstrate the basics of inverting for sparse and/or blocky models. Here, we used the iteratively reweighted least-squares approach. For this tutorial, we focus on the following:
Defining the forward problem
Defining the inverse problem (data misfit, regularization, optimization)
Defining the paramters for the IRLS algorithm
Specifying directives for the inversion
Recovering a set of model parameters which explains the observations
import numpy as np
import matplotlib.pyplot as plt
from discretize import TensorMesh
from SimPEG.simulation import LinearSimulation
from SimPEG.data import Data
from SimPEG import (
simulation,
maps,
data_misfit,
directives,
optimization,
regularization,
inverse_problem,
inversion,
)
# sphinx_gallery_thumbnail_number = 3
Defining the Model and Mapping#
Here we generate a synthetic model and a mappig which goes from the model space to the row space of our linear operator.
nParam = 100 # Number of model paramters
# A 1D mesh is used to define the row-space of the linear operator.
mesh = TensorMesh([nParam])
# Creating the true model
true_model = np.zeros(mesh.nC)
true_model[mesh.cell_centers_x > 0.3] = 1.0
true_model[mesh.cell_centers_x > 0.45] = -0.5
true_model[mesh.cell_centers_x > 0.6] = 0
# Mapping from the model space to the row space of the linear operator
model_map = maps.IdentityMap(mesh)
# Plotting the true model
fig = plt.figure(figsize=(8, 5))
ax = fig.add_subplot(111)
ax.plot(mesh.cell_centers_x, true_model, "b-")
ax.set_ylim([-2, 2])
(-2.0, 2.0)
Defining the Linear Operator#
Here we define the linear operator with dimensions (nData, nParam). In practive, you may have a problem-specific linear operator which you would like to construct or load here.
# Number of data observations (rows)
nData = 20
# Create the linear operator for the tutorial. The columns of the linear operator
# represents a set of decaying and oscillating functions.
jk = np.linspace(1.0, 60.0, nData)
p = -0.25
q = 0.25
def g(k):
return np.exp(p * jk[k] * mesh.cell_centers_x) * np.cos(
np.pi * q * jk[k] * mesh.cell_centers_x
)
G = np.empty((nData, nParam))
for i in range(nData):
G[i, :] = g(i)
# Plot the columns of G
fig = plt.figure(figsize=(8, 5))
ax = fig.add_subplot(111)
for i in range(G.shape[0]):
ax.plot(G[i, :])
ax.set_title("Columns of matrix G")
Text(0.5, 1.0, 'Columns of matrix G')
Defining the Simulation#
The simulation defines the relationship between the model parameters and predicted data.
sim = simulation.LinearSimulation(mesh, G=G, model_map=model_map)
Predict Synthetic Data#
Here, we use the true model to create synthetic data which we will subsequently invert.
# Standard deviation of Gaussian noise being added
std = 0.02
np.random.seed(1)
# Create a SimPEG data object
data_obj = sim.make_synthetic_data(true_model, noise_floor=std, add_noise=True)
Define the Inverse Problem#
The inverse problem is defined by 3 things:
Data Misfit: a measure of how well our recovered model explains the field data
Regularization: constraints placed on the recovered model and a priori information
Optimization: the numerical approach used to solve the inverse problem
# Define the data misfit. Here the data misfit is the L2 norm of the weighted
# residual between the observed data and the data predicted for a given model.
# Within the data misfit, the residual between predicted and observed data are
# normalized by the data's standard deviation.
dmis = data_misfit.L2DataMisfit(simulation=sim, data=data_obj)
# Define the regularization (model objective function). Here, 'p' defines the
# the norm of the smallness term and 'q' defines the norm of the smoothness
# term.
reg = regularization.Sparse(mesh, mapping=model_map)
reg.reference_model = np.zeros(nParam)
p = 0.0
q = 0.0
reg.norms = [p, q]
# Define how the optimization problem is solved.
opt = optimization.ProjectedGNCG(
maxIter=100, lower=-2.0, upper=2.0, maxIterLS=20, maxIterCG=30, tolCG=1e-4
)
# Here we define the inverse problem that is to be solved
inv_prob = inverse_problem.BaseInvProblem(dmis, reg, opt)
Define Inversion Directives#
Here we define any directiveas that are carried out during the inversion. This includes the cooling schedule for the trade-off parameter (beta), stopping criteria for the inversion and saving inversion results at each iteration.
# Add sensitivity weights but don't update at each beta
sensitivity_weights = directives.UpdateSensitivityWeights(everyIter=False)
# Reach target misfit for L2 solution, then use IRLS until model stops changing.
IRLS = directives.Update_IRLS(max_irls_iterations=40, minGNiter=1, f_min_change=1e-4)
# Defining a starting value for the trade-off parameter (beta) between the data
# misfit and the regularization.
starting_beta = directives.BetaEstimate_ByEig(beta0_ratio=1e0)
# Update the preconditionner
update_Jacobi = directives.UpdatePreconditioner()
# Save output at each iteration
saveDict = directives.SaveOutputEveryIteration(save_txt=False)
# Define the directives as a list
directives_list = [sensitivity_weights, IRLS, starting_beta, update_Jacobi, saveDict]
Setting a Starting Model and Running the Inversion#
To define the inversion object, we need to define the inversion problem and the set of directives. We can then run the inversion.
# Here we combine the inverse problem and the set of directives
inv = inversion.BaseInversion(inv_prob, directives_list)
# Starting model
starting_model = 1e-4 * np.ones(nParam)
# Run inversion
recovered_model = inv.run(starting_model)
SimPEG.InvProblem is setting bfgsH0 to the inverse of the eval2Deriv.
***Done using the default solver Pardiso and no solver_opts.***
model has any nan: 0
=============================== Projected GNCG ===============================
# beta phi_d phi_m f |proj(x-g)-x| LS Comment
-----------------------------------------------------------------------------
x0 has any nan: 0
0 1.67e+06 1.88e+03 5.03e-10 1.88e+03 1.99e+01 0
1 8.34e+05 9.46e+02 1.99e-04 1.11e+03 1.89e+01 0
2 4.17e+05 6.46e+02 4.59e-04 8.37e+02 1.82e+01 0 Skip BFGS
3 2.08e+05 3.78e+02 9.16e-04 5.69e+02 1.66e+01 0 Skip BFGS
4 1.04e+05 1.89e+02 1.55e-03 3.51e+02 1.44e+01 0 Skip BFGS
5 5.21e+04 8.29e+01 2.26e-03 2.01e+02 1.24e+01 0 Skip BFGS
6 2.61e+04 3.40e+01 2.90e-03 1.10e+02 9.80e+00 0 Skip BFGS
7 1.30e+04 1.52e+01 3.39e-03 5.94e+01 8.30e+00 0 Skip BFGS
Reached starting chifact with l2-norm regularization: Start IRLS steps...
irls_threshold 1.2879993543552963
8 6.51e+03 8.84e+00 5.11e-03 4.21e+01 3.12e+00 0 Skip BFGS
9 1.05e+04 8.10e+00 5.88e-03 7.01e+01 1.43e+01 0
10 8.20e+03 1.24e+01 5.91e-03 6.09e+01 1.70e+00 0
11 6.39e+03 1.24e+01 6.33e-03 5.28e+01 1.84e+00 0 Skip BFGS
12 5.05e+03 1.20e+01 6.60e-03 4.54e+01 2.47e+00 0 Skip BFGS
13 4.09e+03 1.15e+01 6.67e-03 3.88e+01 2.95e+00 0 Skip BFGS
14 4.09e+03 1.07e+01 6.53e-03 3.74e+01 2.92e+00 0
15 4.09e+03 1.09e+01 6.09e-03 3.58e+01 3.14e+00 0
16 4.09e+03 1.09e+01 5.58e-03 3.37e+01 3.25e+00 0
17 4.09e+03 1.08e+01 5.04e-03 3.15e+01 3.33e+00 0
18 4.09e+03 1.06e+01 4.52e-03 2.91e+01 3.54e+00 0
19 4.09e+03 1.04e+01 3.98e-03 2.67e+01 3.90e+00 0
20 4.09e+03 1.00e+01 3.43e-03 2.41e+01 4.28e+00 0
21 4.09e+03 9.57e+00 2.92e-03 2.15e+01 4.88e+00 0
22 6.37e+03 9.00e+00 2.44e-03 2.45e+01 1.13e+01 0
23 6.37e+03 9.42e+00 1.93e-03 2.17e+01 5.05e+00 0
24 6.37e+03 9.00e+00 1.63e-03 1.94e+01 5.05e+00 0 Skip BFGS
25 1.00e+04 8.66e+00 1.50e-03 2.37e+01 1.37e+01 0 Skip BFGS
26 1.00e+04 9.33e+00 1.20e-03 2.14e+01 7.66e+00 0
27 1.00e+04 9.08e+00 1.04e-03 1.95e+01 8.84e+00 0
28 1.58e+04 8.78e+00 8.86e-04 2.27e+01 1.53e+01 0
29 1.58e+04 9.15e+00 7.13e-04 2.04e+01 9.34e+00 0
30 1.58e+04 9.07e+00 5.93e-04 1.84e+01 9.05e+00 0 Skip BFGS
31 2.46e+04 8.95e+00 4.99e-04 2.12e+01 1.70e+01 0
32 2.46e+04 9.21e+00 4.04e-04 1.91e+01 9.95e+00 0
33 2.46e+04 9.11e+00 3.29e-04 1.72e+01 9.58e+00 0
34 3.83e+04 8.97e+00 2.66e-04 1.92e+01 1.69e+01 0
35 3.83e+04 9.12e+00 2.14e-04 1.73e+01 9.82e+00 0
36 3.83e+04 9.02e+00 1.78e-04 1.58e+01 9.34e+00 0 Skip BFGS
37 5.96e+04 8.97e+00 1.49e-04 1.78e+01 1.73e+01 0
38 5.96e+04 9.02e+00 1.22e-04 1.63e+01 9.57e+00 0
39 9.28e+04 8.97e+00 9.91e-05 1.82e+01 1.63e+01 0
40 9.28e+04 9.07e+00 8.12e-05 1.66e+01 9.34e+00 0
41 9.28e+04 9.05e+00 6.77e-05 1.53e+01 1.11e+01 0 Skip BFGS
42 1.44e+05 8.99e+00 5.64e-05 1.71e+01 1.59e+01 0
43 1.44e+05 9.05e+00 4.68e-05 1.58e+01 9.26e+00 0
44 1.44e+05 9.06e+00 3.89e-05 1.47e+01 9.22e+00 0 Skip BFGS
45 1.44e+05 9.06e+00 3.24e-05 1.37e+01 9.18e+00 0 Skip BFGS
46 1.44e+05 9.04e+00 2.70e-05 1.29e+01 9.16e+00 0 Skip BFGS
47 1.44e+05 9.03e+00 2.25e-05 1.23e+01 9.14e+00 0 Skip BFGS
Reach maximum number of IRLS cycles: 40
------------------------- STOP! -------------------------
1 : |fc-fOld| = 0.0000e+00 <= tolF*(1+|f0|) = 1.8853e+02
1 : |xc-x_last| = 6.2523e-02 <= tolX*(1+|x0|) = 1.0010e-01
0 : |proj(x-g)-x| = 9.1422e+00 <= tolG = 1.0000e-01
0 : |proj(x-g)-x| = 9.1422e+00 <= 1e3*eps = 1.0000e-02
0 : maxIter = 100 <= iter = 48
------------------------- DONE! -------------------------
Plotting Results#
fig, ax = plt.subplots(1, 2, figsize=(12 * 1.2, 4 * 1.2))
# True versus recovered model
ax[0].plot(mesh.cell_centers_x, true_model, "k-")
ax[0].plot(mesh.cell_centers_x, inv_prob.l2model, "b-")
ax[0].plot(mesh.cell_centers_x, recovered_model, "r-")
ax[0].legend(("True Model", "Recovered L2 Model", "Recovered Sparse Model"))
ax[0].set_ylim([-2, 2])
# Observed versus predicted data
ax[1].plot(data_obj.dobs, "k-")
ax[1].plot(inv_prob.dpred, "ko")
ax[1].legend(("Observed Data", "Predicted Data"))
# Plot convergence
fig = plt.figure(figsize=(9, 5))
ax = fig.add_axes([0.2, 0.1, 0.7, 0.85])
ax.plot(saveDict.phi_d, "k", lw=2)
twin = ax.twinx()
twin.plot(saveDict.phi_m, "k--", lw=2)
ax.plot(np.r_[IRLS.iterStart, IRLS.iterStart], np.r_[0, np.max(saveDict.phi_d)], "k:")
ax.text(
IRLS.iterStart,
0.0,
"IRLS Start",
va="bottom",
ha="center",
rotation="vertical",
size=12,
bbox={"facecolor": "white"},
)
ax.set_ylabel("$\phi_d$", size=16, rotation=0)
ax.set_xlabel("Iterations", size=14)
twin.set_ylabel("$\phi_m$", size=16, rotation=0)
Text(865.1527777777777, 0.5, '$\\phi_m$')
Total running time of the script: ( 0 minutes 30.071 seconds)
Estimated memory usage: 18 MB