Geophysical inversion frequently makes use of regularization, such as the “Tikhonov regularization” used by Kenneth Bube and Bob Langan  for their “continuation approach.” I'd like to suggest an adjustment of the objective function to allow faster convergence of regularization and the continuation approach. A damping term that discourages complexity can be replaced equivalently by a change of variables to model simplicity directly.
For an optimized inversion, an objective function typically includes a norm of the difference between a data vector and a non-linear transform of a model vector . The global minimum of this norm is often flat, with little sensitivity to large variations in the model.
For regularization (more than simple damping), a linear operator is chosen to remove simplicity and preserve complexity when applied to the model vector as . Most examples use a roughening operator, such as a derivative, to suppress long wavelengths and amplify short wavelengths. A regularized objective function adds a norm of this roughened model to the norm fitting the data:objective function is easily motivated as a maximum a posteriori estimate of the model given the data. Additive noise is assumed to be Gaussian and uncorrelated with zero mean. The model is assumed to be Gaussian and zero mean, with an inverse covariance matrix equal to adjusts the ratio of variances assumed for noise and the model.
Bube and Langan's continuation approach begins with a large constant , minimizes the objective function (1) for a first model, then reduces repeatedly for a tradeoff between simplicity and accuracy in fitting the recorded data. They find the simplest model possible to explain the data adequately, without preventing the model from using complexity to fit genuinely significant features of the data. Informative details are added to the model when justified by the data, without unnecessary distracting details that are poorly determined from the data.
Each minimization of the objective function (1)
for a fixed constant typically uses a descent method such
as Gauss-Newton with conjugate gradients. The properties of the gradient
are important to the rate of convergence:
Assume a new operator as a partial right inverse of , so that the two operators approximate an identity: . This operator should be designed to preserve simplicity and suppress complexity, although without destroying complexity entirely. If is a roughening operator like differentiation, then should be a smoothing operator like leaky integration.
More directly, define the simplification operator as a factored form of the assumed covariance. (Indeed, such a factorization always exists because the covariance is positive semidefinite.)
Minimization of the original objective function (1) is entirely equivalent to minimizing the objective function with a new variable , where :now has uncorrelated samples. Although we optimize this new model , we keep and use the original model . Continuation can adjust the constant as before, with identical results (assuming complete minimization of the objective functions  and ).
The revised gradient contains the desired correlation:
The two objective functions produce different results when optimization is incomplete. A descent optimization of the original objective function (1) will begin with complex perturbations of the model and slowly converge toward an increasingly simple model at the global minimum. A descent optimization of the revised objective function (6) will begin with simple perturbations of the model and slowly converge toward an increasingly complex model at the global minimum. The latter strategy is more consistent with the overall goal of the continuation approach. A more economic implementation can use fewer iterations. Insufficient iterations result in an insufficiently complex model, not in an insufficiently simplified model.
I also prefer to adjust more than a single scale factor . Instead, assume a suite of simplification operators which allow increasing complexity as the index increases. (Furthermore and .) We then can optimize a suite of possible models, of increasing complexity as increases. Use each optimized model to initialize the next . As multigrid methods have shown, we can thus improve our overall convergence by optimizing the most reliable (smoothest) global features in the model before attempting finer detail.
Finally, I think it easier to choose a simplification operator which describes the desirable features of the model, rather than an operator which keeps only features thought to be undesirable. I see some value in constructing both, however, to check the consistency of assumptions.