Abstract

SUMMARYIn probabilistic Bayesian inversions, data uncertainty is a crucial parameter for quantifying the uncertainties and correlations of the resulting model parameters or, in transdimensional approaches, even the complexity of the model. However, in many geophysical inference problems it is poorly known. Therefore, it is common practice to allow the data uncertainty itself to be a parameter to be determined. Although in principle any arbitrary uncertainty distribution can be assumed, Gaussian distributions whose standard deviation is then the unknown parameter to be estimated are the usual choice. In this special case, the paper demonstrates that a simple analytical integration is sufficient to marginalise out this uncertainty parameter, reducing the complexity of the model space without compromising the accuracy of the posterior model probability distribution. However, it is well known that the distribution of geophysical measurement errors, although superficially similar to a Gaussian distribution, typically contains more frequent samples along the tail of the distribution, so-called outliers. In linearized inversions these are often removed in subsequent iterations based on some threshold criterion, but in Markov chain Monte Carlo (McMC) inversions this approach is not possible as they rely on the likelihood ratios, which cannot be formed if the number of data points varies between the steps of the Markov chain. The flexibility to define the data error probability distribution in McMC can be exploited in order to account for this pattern of uncertainties in a natural way, without having to make arbitrary choices regarding residual thresholds. In particular, we can regard the data uncertainty distribution as a mixture between a Gaussian distribution, which represent valid measurements with some measurement error, and a uniform distribution, which represents invalid measurements. The relative balance between them is an unknown parameter to be estimated alongside the standard deviation of the Gauss distribution. For each data point, the algorithm can then assign a probability to be an outlier, and the influence of each data point will be effectively downgraded according to its probability to be an outlier. Furthermore, this assignment can change as the McMC search is exploring different parts of the model space. The approach is demonstrated with both synthetic and real tomography examples. In a synthetic test, the proposed mixed measurement error distribution allows recovery of the underlying model even in the presence of 6 per cent outliers, which completely destroy the ability of a regular McMC or linear search to provide a meaningful image. Applied to an actual ambient noise tomography study based on automatically picked dispersion curves, the resulting model is shown to be much more consistent for different data sets, which differ in the applied quality criteria, while retaining the ability to recover strong anomalies in selected parts of the model.

Highlights

  • In the past, solving a geophysical inverse problem generally implied finding an optimum model that fits the observed data in a least squares sense(e.g., Backus & F. 1967; Menke 2012; Aster et al 2005) and fulfils a number of essentially arbitrary regularisation constraints (Engl et al 2000) such as damping or smoothing

  • SUMMARY In probabilistic Bayesian inversions, data uncertainty is a crucial parameter for quantifying the uncertainties and correlations of the resulting model parameters or, in transdimensional approaches, even the complexity of the model

  • The paper demonstrates that a simple analytical integration is sufficient to marginalise out this uncertainty parameter, reducing the complexity of the model space without compromising the accuracy of the posterior model probability distribution

Read more

Summary

INTRODUCTION

In the past, solving a geophysical inverse problem generally implied finding an optimum model that fits the observed data in a least squares sense(e.g., Backus & F. 1967; Menke 2012; Aster et al 2005) and fulfils a number of essentially arbitrary regularisation constraints (Engl et al 2000) such as damping (minimisation of model derivatives) or smoothing (minimisation of first or second order derivatives). 1967; Menke 2012; Aster et al 2005) and fulfils a number of essentially arbitrary regularisation constraints (Engl et al 2000) such as damping (minimisation of model derivatives) or smoothing (minimisation of first or second order derivatives) Whereas this paradigm persists for computing-intensive inverse problems such as full-waveform modelling in two or three dimensions (e.g., Komatitsch & Vilotte 1998; Virieux & Operto 2009), increasing computer power has enabled the practical application of algorithms that apply Bayes’ theorem (Bayes 1763) to generate an ‘optimum model’ but an estimate of the probability distribution of the model parameters, m, given the observed data d, and an a priori probability distribution of the model parameters, p(m). The second part of the paper will introduce a mixed probability distribution which is explicitly accounting for outliers

GAUSSIAN DATA UNCERTAINTY
Application to mean value estimation
Application to synthetic tomography problem
Findings
USE CASE
DISCUSSION AND CONCLUSION

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.