Network inference via approximate Bayesian computation. Illustration on a stochastic multipopulation neural mass model
Network inference via approximate Bayesian computation. Illustration on a stochastic multipopulation neural mass model
- Research Article
14
- 10.1016/j.neuroimage.2021.118020
- Apr 9, 2021
- Neuroimage
This paper describes and validates a novel framework using the Approximate Bayesian Computation (ABC) algorithm for parameter estimation and model selection in models of mesoscale brain network activity. We provide a proof of principle, first pass validation of this framework using a set of neural mass models of the cortico-basal ganglia thalamic circuit inverted upon spectral features from experimental, in vivo recordings. This optimization scheme relaxes an assumption of fixed-form posteriors (i.e. the Laplace approximation) taken in previous approaches to inverse modelling of spectral features. This enables the exploration of model dynamics beyond that approximated from local linearity assumptions and so fit to explicit, numerical solutions of the underlying non-linear system of equations. In this first paper, we establish a face validation of the optimization procedures in terms of: (i) the ability to approximate posterior densities over parameters that are plausible given the known causes of the data; (ii) the ability of the model comparison procedures to yield posterior model probabilities that can identify the model structure known to generate the data; and (iii) the robustness of these procedures to local minima in the face of different starting conditions. Finally, as an illustrative application we show (iv) that model comparison can yield plausible conclusions given the known neurobiology of the cortico-basal ganglia-thalamic circuit in Parkinsonism. These results lay the groundwork for future studies utilizing highly nonlinear or brittle models that can explain time dependant dynamics, such as oscillatory bursts, in terms of the underlying neural circuits.
- Research Article
- 10.3389/conf.fninf.2016.20.00014
- Jan 1, 2016
- Frontiers in Neuroinformatics
Event Abstract Back to Event Statistical uncertainty and sensitivity analysis of intracellular signaling models - through approximate Bayesian computation and variance based global sensitivity analysis Olivia Eriksson1, 2*, Alexandra Jauhiainen1, 3, Sara Maad Sasane4, Anu G Nair5, Carolina Sartorius4 and Jeanette Hellgren Kotaleski2, 5, 6 1 Contributed equally to this work, Sweden 2 Stockholm University, Department of Numerical Analysis and Computer Science, Sweden 3 AstraZeneca AB R&D, Early Clinical Biometrics, Sweden 4 Lund University, Centre for Mathematical Sciences, Sweden 5 KTH Royal Institute of Technology, School of Computer Science and Communication, Sweden 6 Karolinska Institute, Department of Neuroscience, Sweden Computational dynamical models describing intracellular signaling pathways of nerve cells are increasing in size and complexity as more information is obtained from experiments. Models describing protein interactions within neurons can contain over hundred protein reactions [1, 2]. These models are built from knowledge about the interaction topology, inferred from e.g. gene knock-out experiments, as well as from experimental quantitative data describing the input-output relationship of the observed system. The quantitative data are often sparse as compared to the size of the system, and translating the experimental information into dynamical models often results in large uncertainties in the parameter (e.g reaction rates) values. This uncertainty is partly due to the sparseness in experimental time series data (practical unidentifiability), but it is also an intrinsic feature of the modeled system (structural unidentifiability) and could reflect the possibility for biological variation [3]; in many biological systems the same function can be achieved in partly different ways due to degeneracy and redundancy. Here we investigate the extent of this parameter uncertainty, in a model describing calcium (Ca)-dependent activation of Calmodulin (CaM), Protein phosphatase 2B (PP2B) and Ca/CaM-dependent protein kinase II (CamKII) [4], which is an intracellular pathway of importance to synaptic plasticity. We presume that the structure of the model is correct and then characterize the uncertainty in the parameters based on data. This is done by sampling the subspace of parameter values which has an equally good fit to data (corresponding to the “viable space” in [5]), through a statistical approach known as Approximate Bayesian Computation (ABC) with Markov Chain Monte Carlo (MCMC) [6]. ABC has been proposed for parameter estimation in system biology models previously, albeit in slightly different settings, e.g. [7, 8], and can be applied to a wide range of problems. In order to characterize the viable space efficiently with several experimental datasets, we employ multivariate probability distributions called copulas as a part of the ABC sampling. The result is a multivariate posterior distribution for the parameters, which represents how much they are constrained by the data (i.e. the uncertainty) as illustrated in Figure 1 (pairwise plots of viable space for four parameters indicating constraints and dependencies before and after calibration to data). The next step is to look into how the parameter uncertainty is translated to uncertainty of the predictions made from the model. We also investigate the role and importance of different parameters, with respect to different model outputs including the predictions, by means of global sensitivity analysis [9]. Figure 2 illustrates the sensitivity of one of the outputs calculated by a method based on the decomposition of the variance [10, 11, 12], showing different types of sensitivity measures. Finally, we explore how the sensitivity of different subsets of parameters change as different parts of the viable space are investigated, which is shown in Figure 3, where sensitivity profiles are clustered into groups with different sensitivity characteristics. Figure 1 Figure 2 Figure 3 Acknowledgements This research was supported by funding from the Swedish e-Science Research Centre and the Human Brain Project References [1] Gutierrez-Arenas O, Eriksson O, Hellgren Kotaleski J (2014) Segregation and Crosstalk of D1 Receptor-Mediated Activation of ERK in Striatal Medium Spiny Neurons upon Acute Administration of Psychostimulants. PLoS Comput Biol 10, e1003445. [2] Nair AG, Gutierrez-Arenas O, Eriksson O, Vincent P, and Hellgren Kotaleski J (2015) Sensing Positive versus Negative Reward Signals through Adenylyl Cyclase-Coupled GPCRs in Direct and Indirect Pathway Striatal Medium Spiny Neurons. J. Neuroscience 35, 14017-14030. [3] Marder E, Goeritz ML, Otopalik AG (2015) Robust circuit rhythms in small circuits arise from variable circuit components and mechanisms. Curr Opin Neurobiol 31, 156-163. [4] Nair AG, Gutierrez-Arenas O, Eriksson O, Jauhiainen A, Blackwell KT and Kotaleski, JH (2014). Modeling intracellular signaling underlying striatal function in health and disease. Progress in molecular biology and translational science, 123, 277 [5] Zamora-Sillero E, Hafner M, Ibig A, Stelling J, and Wagner, A. (2011). Efficient characterization of high-dimensional parameter spaces for systems biology. BMC systems biology 5, 142. [6] Marjoram P, Molitor J, Plagnol V, and Tavare S (2003) Markov chain Monte Carlo without likelihoods. Proc. Natl. Acad. Sci. U.S.A. 100, 15324–15328. [7] Toni T, Welch D, Strelkowa N, Ipsen A, and Stumpf MP. Approximate Bayesian computation scheme for parameter inference and model selection in dynamical systems. J R Soc Interface, 6, 187–202. [8] Liepe J, Kirk P, Filippi S, Toni T, Barnes CP, and Stumpf MP. A framework for parameter estimation and model selection from experimental data in systems biology using approximate Bayesian computation. Nat Protoc, 9, 439–456. [9] Saltelli A, Ratto M, Andres T, Campolongo F, Cariboni J, Gatelli D, Saisana M, Tarantola S Global sensitivity analysis: the primer. In Wiley 2008 New York, NY:Wiley. [10] Sobol IM (2001) Global sensitivity indices for nonlinear mathematical models and their Monte Carlo estimates, Math. Comput. Simul. 55, 271–280. [11] Saltelli A (2002) Making best use of model evaluations to compute sensitivity indices. Comp Phys Comm 145, 280 – 297. [12] Halnes G, Ulfhielm E Ljunggren EE, Kotaleski JH and Rospars JP (2009) Modelling and sensitivity analysis of the reactions involving receptor, G-protein and effector in vertebrate olfactory receptor neurons. Journal of computational neuroscience 27, 471-491. Keywords: sensitivity analysis, Approximate Bayesian Computation, copulas, intracellular signalling, Dynamical Modeling, global sensitivity analysis, Markov chain Monte Carlo, variance decomposition Conference: Neuroinformatics 2016, Reading, United Kingdom, 3 Sep - 4 Sep, 2016. Presentation Type: Investigator presentations Topic: Computational neuroscience Citation: Eriksson O, Jauhiainen A, Maad Sasane S, Nair A, Sartorius C and Hellgren Kotaleski J (2016). Statistical uncertainty and sensitivity analysis of intracellular signaling models - through approximate Bayesian computation and variance based global sensitivity analysis. Front. Neuroinform. Conference Abstract: Neuroinformatics 2016. doi: 10.3389/conf.fninf.2016.20.00014 Copyright: The abstracts in this collection have not been subject to any Frontiers peer review or checks, and are not endorsed by Frontiers. They are made available through the Frontiers publishing platform as a service to conference organizers and presenters. The copyright in the individual abstracts is owned by the author of each abstract or his/her employer unless otherwise stated. Each abstract, as well as the collection of abstracts, are published under a Creative Commons CC-BY 4.0 (attribution) licence (https://creativecommons.org/licenses/by/4.0/) and may thus be reproduced, translated, adapted and be the subject of derivative works provided the authors and Frontiers are attributed. For Frontiers’ terms and conditions please see https://www.frontiersin.org/legal/terms-and-conditions. Received: 31 May 2016; Published Online: 18 Jul 2016. * Correspondence: PhD. Olivia Eriksson, Contributed equally to this work, Stockholm, Sweden, olivia@kth.se Login Required This action requires you to be registered with Frontiers and logged in. To register or login click here. Abstract Info Abstract The Authors in Frontiers Olivia Eriksson Alexandra Jauhiainen Sara Maad Sasane Anu G Nair Carolina Sartorius Jeanette Hellgren Kotaleski Google Olivia Eriksson Alexandra Jauhiainen Sara Maad Sasane Anu G Nair Carolina Sartorius Jeanette Hellgren Kotaleski Google Scholar Olivia Eriksson Alexandra Jauhiainen Sara Maad Sasane Anu G Nair Carolina Sartorius Jeanette Hellgren Kotaleski PubMed Olivia Eriksson Alexandra Jauhiainen Sara Maad Sasane Anu G Nair Carolina Sartorius Jeanette Hellgren Kotaleski Related Article in Frontiers Google Scholar PubMed Abstract Close Back to top Javascript is disabled. Please enable Javascript in your browser settings in order to see all the content on this page.
- Research Article
110
- 10.1111/j.1365-294x.2008.04026.x
- Jan 1, 2009
- Molecular ecology
Nested clade phylogeographical analysis (NCPA) and approximate Bayesian computation (ABC) have been used to test phylogeographical hypotheses. Multilocus NCPA tests null hypotheses, whereas ABC discriminates among a finite set of alternatives. The interpretive criteria of NCPA are explicit and allow complex models to be built from simple components. The interpretive criteria of ABC are ad hoc and require the specification of a complete phylogeographical model. The conclusions from ABC are often influenced by implicit assumptions arising from the many parameters needed to specify a complex model. These complex models confound many assumptions so that biological interpretations are difficult. Sampling error is accounted for in NCPA, but ABC ignores important sources of sampling error that creates pseudo-statistical power. NCPA generates the full sampling distribution of its statistics, but ABC only yields local probabilities, which in turn make it impossible to distinguish between a good fitting model, a non-informative model, and an over-determined model. Both NCPA and ABC use approximations, but convergences of the approximations used in NCPA are well defined whereas those in ABC are not. NCPA can analyse a large number of locations, but ABC cannot. Finally, the dimensionality of tested hypothesis is known in NCPA, but not for ABC. As a consequence, the 'probabilities' generated by ABC are not true probabilities and are statistically non-interpretable. Accordingly, ABC should not be used for hypothesis testing, but simulation approaches are valuable when used in conjunction with NCPA or other methods that do not rely on highly parameterized models.
- Research Article
48
- 10.1093/biomet/asx078
- Jan 20, 2018
- Biometrika
Many statistical applications involve models for which it is difficult to evaluate the likelihood, but from which it is relatively easy to sample. Approximate Bayesian computation is a likelihood-free method for implementing Bayesian inference in such cases. We present results on the asymptotic variance of estimators obtained using approximate Bayesian computation in a large-data limit. Our key assumption is that the data are summarized by a fixed-dimensional summary statistic that obeys a central limit theorem. We prove asymptotic normality of the mean of the approximate Bayesian computation posterior. This result also shows that, in terms of asymptotic variance, we should use a summary statistic that is the same dimension as the parameter vector, p; and that any summary statistic of higher dimension can be reduced, through a linear transformation, to dimension p in a way that can only reduce the asymptotic variance of the posterior mean. We look at how the Monte Carlo error of an importance sampling algorithm that samples from the approximate Bayesian computation posterior affects the accuracy of estimators. We give conditions on the importance sampling proposal distribution such that the variance of the estimator will be the same order as that of the maximum likelihood estimator based on the summary statistics used. This suggests an iterative importance sampling algorithm, which we evaluate empirically on a stochastic volatility model.
- Research Article
3
- 10.1214/20-ba1248
- Dec 9, 2020
- Bayesian analysis
Approximate Bayesian computation (ABC) is a simulation-based likelihood-free method applicable to both model selection and parameter estimation. ABC parameter estimation requires the ability to forward simulate datasets from a candidate model, but because the sizes of the observed and simulated datasets usually need to match, this can be computationally expensive. Additionally, since ABC inference is based on comparisons of summary statistics computed on the observed and simulated data, using computationally expensive summary statistics can lead to further losses in efficiency. ABC has recently been applied to the family of mechanistic network models, an area that has traditionally lacked tools for inference and model choice. Mechanistic models of network growth repeatedly add nodes to a network until it reaches the size of the observed network, which may be of the order of millions of nodes. With ABC, this process can quickly become computationally prohibitive due to the resource intensive nature of network simulations and evaluation of summary statistics. We propose two methodological developments to enable the use of ABC for inference in models for large growing networks. First, to save time needed for forward simulating model realizations, we propose a procedure to extrapolate (via both least squares and Gaussian processes) summary statistics from small to large networks. Second, to reduce computation time for evaluating summary statistics, we use sample-based rather than census-based summary statistics. We show that the ABC posterior obtained through this approach, which adds two additional layers of approximation to the standard ABC, is similar to a classic ABC posterior. Although we deal with growing network models, both extrapolated summaries and sampled summaries are expected to be relevant in other ABC settings where the data are generated incrementally.
- Research Article
113
- 10.1016/j.ecolmodel.2015.05.020
- Jun 11, 2015
- Ecological Modelling
This paper investigates the feasibility of using Approximate Bayesian Computation (ABC) to calibrate and evaluate complex individual-based models (IBMs). As ABC evolves, various versions are emerging, but here we only explore the most accessible version, rejection-ABC. Rejection-ABC involves running models a large number of times, with parameters drawn randomly from their prior distributions, and then retaining the simulations closest to the observations. Although well-established in some fields, whether ABC will work with ecological IBMs is still uncertain.Rejection-ABC was applied to an existing 14-parameter earthworm energy budget IBM for which the available data consist of body mass growth and cocoon production in four experiments. ABC was able to narrow the posterior distributions of seven parameters, estimating credible intervals for each. ABC's accepted values produced slightly better fits than literature values do. The accuracy of the analysis was assessed using cross-validation and coverage, currently the best-available tests. Of the seven unnarrowed parameters, ABC revealed that three were correlated with other parameters, while the remaining four were found to be not estimable given the data available.It is often desirable to compare models to see whether all component modules are necessary. Here, we used ABC model selection to compare the full model with a simplified version which removed the earthworm's movement and much of the energy budget. We are able to show that inclusion of the energy budget is necessary for a good fit to the data. We show how our methodology can inform future modelling cycles, and briefly discuss how more advanced versions of ABC may be applicable to IBMs. We conclude that ABC has the potential to represent uncertainty in model structure, parameters and predictions, and to embed the often complex process of optimising an IBM's structure and parameters within an established statistical framework, thereby making the process more transparent and objective.
- Research Article
6
- 10.1016/j.ecolmodel.2022.110251
- Dec 26, 2022
- Ecological Modelling
Calibration of a bumble bee foraging model using Approximate Bayesian Computation
- Research Article
184
- 10.1093/sysbio/syw077
- Sep 11, 2016
- Systematic Biology
Bayesian inference plays an important role in phylogenetics, evolutionary biology, and in many other branches of science. It provides a principled framework for dealing with uncertainty and quantifying how it changes in the light of new evidence. For many complex models and inference problems, however, only approximate quantitative answers are obtainable. Approximate Bayesian computation (ABC) refers to a family of algorithms for approximate inference that makes a minimal set of assumptions by only requiring that sampling from a model is possible. We explain here the fundamentals of ABC, review the classical algorithms, and highlight recent developments. [ABC; approximate Bayesian computation; Bayesian inference; likelihood-free inference; phylogenetics; simulator-based models; stochastic simulation models; tree-based models.]
- Research Article
30
- 10.1007/s11222-013-9432-2
- Nov 29, 2013
- Statistics and Computing
Many modern statistical applications involve inference for complicated stochastic models for which the likelihood function is difficult or even impossible to calculate, and hence conventional likelihood-based inferential techniques cannot be used. In such settings, Bayesian inference can be performed using Approximate Bayesian Computation (ABC). However, in spite of many recent developments to ABC methodology, in many applications the computational cost of ABC necessitates the choice of summary statistics and tolerances that can potentially severely bias the estimate of the posterior.We propose a new “piecewise” ABC approach suitable for discretely observed Markov models that involves writing the posterior density of the parameters as a product of factors, each a function of only a subset of the data, and then using ABC within each factor. The approach has the advantage of side-stepping the need to choose a summary statistic and it enables a stringent tolerance to be set, making the posterior “less approximate”. We investigate two methods for estimating the posterior density based on ABC samples for each of the factors: the first is to use a Gaussian approximation for each factor, and the second is to use a kernel density estimate. Both methods have their merits. The Gaussian approximation is simple, fast, and probably adequate for many applications. On the other hand, using instead a kernel density estimate has the benefit of consistently estimating the true piecewise ABC posterior as the number of ABC samples tends to infinity. We illustrate the piecewise ABC approach with four examples; in each case, the approach offers fast and accurate inference.
- Research Article
29
- 10.1155/2013/210646
- Jan 1, 2013
- BioMed Research International
Approximate Bayesian computation (ABC) is an approach for using measurement data to calibrate stochastic computer models, which are common in biology applications. ABC is becoming the “go-to” option when the data and/or parameter dimension is large because it relies on user-chosen summary statistics rather than the full data and is therefore computationally feasible. One technical challenge with ABC is that the quality of the approximation to the posterior distribution of model parameters depends on the user-chosen summary statistics. In this paper, the user requirement to choose effective summary statistics in order to accurately estimate the posterior distribution of model parameters is investigated and illustrated by example, using a model and corresponding real data of mitochondrial DNA population dynamics. We show that for some choices of summary statistics, the posterior distribution of model parameters is closely approximated and for other choices of summary statistics, the posterior distribution is not closely approximated. A strategy to choose effective summary statistics is suggested in cases where the stochastic computer model can be run at many trial parameter settings, as in the example.
- Research Article
23
- 10.1007/s11222-019-09909-6
- Nov 5, 2019
- Statistics and Computing
Approximate Bayesian computation (ABC) has become one of the major tools of likelihood-free statistical inference in complex mathematical models. Simultaneously, stochastic differential equations (SDEs) have developed to an established tool for modelling time-dependent, real-world phenomena with underlying random effects. When applying ABC to stochastic models, two major difficulties arise: First, the derivation of effective summary statistics and proper distances is particularly challenging, since simulations from the stochastic process under the same parameter configuration result in different trajectories. Second, exact simulation schemes to generate trajectories from the stochastic model are rarely available, requiring the derivation of suitable numerical methods for the synthetic data generation. To obtain summaries that are less sensitive to the intrinsic stochasticity of the model, we propose to build up the statistical method (e.g. the choice of the summary statistics) on the underlying structural properties of the model. Here, we focus on the existence of an invariant measure and we map the data to their estimated invariant density and invariant spectral density. Then, to ensure that these model properties are kept in the synthetic data generation, we adopt measure-preserving numerical splitting schemes. The derived property-based and measure-preserving ABC method is illustrated on the broad class of partially observed Hamiltonian type SDEs, both with simulated data and with real electroencephalography data. The derived summaries are particularly robust to the model simulation, and this fact, combined with the proposed reliable numerical scheme, yields accurate ABC inference. In contrast, the inference returned using standard numerical methods (Euler–Maruyama discretisation) fails. The proposed ingredients can be incorporated into any type of ABC algorithm and directly applied to all SDEs that are characterised by an invariant distribution and for which a measure-preserving numerical method can be derived.
- Research Article
108
- 10.1016/j.epidem.2019.100368
- Sep 25, 2019
- Epidemics
Approximate Bayesian Computation for infectious disease modelling
- Research Article
51
- 10.1111/rssb.12356
- Jan 8, 2020
- Journal of the Royal Statistical Society Series B: Statistical Methodology
SummaryWe analyse the behaviour of approximate Bayesian computation (ABC) when the model generating the simulated data differs from the actual data-generating process, i.e. when the data simulator in ABC is misspecified. We demonstrate both theoretically and in simple, but practically relevant, examples that when the model is misspecified different versions of ABC can yield substantially different results. Our theoretical results demonstrate that even though the model is misspecified, under regularity conditions, the accept–reject ABC approach concentrates posterior mass on an appropriately defined pseudotrue parameter value. However, under model misspecification the ABC posterior does not yield credible sets with valid frequentist coverage and has non-standard asymptotic behaviour. In addition, we examine the theoretical behaviour of the popular local regression adjustment to ABC under model misspecification and demonstrate that this approach concentrates posterior mass on a pseudotrue value that is completely different from accept–reject ABC. Using our theoretical results, we suggest two approaches to diagnose model misspecification in ABC. All theoretical results and diagnostics are illustrated in a simple running example.
- Research Article
44
- 10.1093/mnras/stx894
- Apr 17, 2017
- Monthly Notices of the Royal Astronomical Society
Standard approaches to Bayesian parameter inference in large scale structure assume a Gaussian functional form (chi-squared form) for the likelihood. This assumption, in detail, cannot be correct. Likelihood free inferences such as Approximate Bayesian Computation (ABC) relax these restrictions and make inference possible without making any assumptions on the likelihood. Instead ABC relies on a forward generative model of the data and a metric for measuring the distance between the model and data. In this work, we demonstrate that ABC is feasible for LSS parameter inference by using it to constrain parameters of the halo occupation distribution (HOD) model for populating dark matter halos with galaxies. Using specific implementation of ABC supplemented with Population Monte Carlo importance sampling, a generative forward model using HOD, and a distance metric based on galaxy number density, two-point correlation function, and galaxy group multiplicity function, we constrain the HOD parameters of mock observation generated from selected "true" HOD parameters. The parameter constraints we obtain from ABC are consistent with the "true" HOD parameters, demonstrating that ABC can be reliably used for parameter inference in LSS. Furthermore, we compare our ABC constraints to constraints we obtain using a pseudo-likelihood function of Gaussian form with MCMC and find consistent HOD parameter constraints. Ultimately our results suggest that ABC can and should be applied in parameter inference for LSS analyses.
- Research Article
80
- 10.1534/genetics.108.098129
- Apr 1, 2009
- Genetics
In recent years approximate Bayesian computation (ABC) methods have become popular in population genetics as an alternative to full-likelihood methods to make inferences under complex demographic models. Most ABC methods rely on the choice of a set of summary statistics to extract information from the data. In this article we tested the use of the full allelic distribution directly in an ABC framework. Although the ABC techniques are becoming more widely used, there is still uncertainty over how they perform in comparison with full-likelihood methods. We thus conducted a simulation study and provide a detailed examination of ABC in comparison with full likelihood in the case of a model of admixture. This model assumes that two parental populations mixed at a certain time in the past, creating a hybrid population, and that the three populations then evolve under pure drift. Several aspects of ABC methodology were investigated, such as the effect of the distance metric chosen to measure the similarity between simulated and observed data sets. Results show that in general ABC provides good approximations to the posterior distributions obtained with the full-likelihood method. This suggests that it is possible to apply ABC using allele frequencies to make inferences in cases where it is difficult to select a set of suitable summary statistics and when the complexity of the model or the size of the data set makes it computationally prohibitive to use full-likelihood methods.
- Ask R Discovery
- Chat PDF
AI summaries and top papers from 250M+ research sources.