Abstract

Approximate Bayesian computation (ABC) is a simulation-based likelihood-free method applicable to both model selection and parameter estimation. ABC parameter estimation requires the ability to forward simulate datasets from a candidate model, but because the sizes of the observed and simulated datasets usually need to match, this can be computationally expensive. Additionally, since ABC inference is based on comparisons of summary statistics computed on the observed and simulated data, using computationally expensive summary statistics can lead to further losses in efficiency. ABC has recently been applied to the family of mechanistic network models, an area that has traditionally lacked tools for inference and model choice. Mechanistic models of network growth repeatedly add nodes to a network until it reaches the size of the observed network, which may be of the order of millions of nodes. With ABC, this process can quickly become computationally prohibitive due to the resource intensive nature of network simulations and evaluation of summary statistics. We propose two methodological developments to enable the use of ABC for inference in models for large growing networks. First, to save time needed for forward simulating model realizations, we propose a procedure to extrapolate (via both least squares and Gaussian processes) summary statistics from small to large networks. Second, to reduce computation time for evaluating summary statistics, we use sample-based rather than census-based summary statistics. We show that the ABC posterior obtained through this approach, which adds two additional layers of approximation to the standard ABC, is similar to a classic ABC posterior. Although we deal with growing network models, both extrapolated summaries and sampled summaries are expected to be relevant in other ABC settings where the data are generated incrementally.

Highlights

  • IntroductionStatistical network models directly model the observed network data

  • An alternative strategy to benefit from the GP marginal distributions is to marginally evaluate each normal density at the observed summary statistic value and use, for example, the sum of these evaluated densities as a similarity measure between observed and extrapolated summary statistics. Because this methodology does not consider the correlation among summary statistics, we focused our investigations on GPa-Approximate Bayesian computation (ABC)

  • We proposed two methodological developments to make ABC feasible for modeling large networks using mechanistic models

Read more

Summary

Introduction

Statistical network models directly model the observed network data Their likelihood functions are usually available in closed form (possibly up to a normalizing constant), and inference and model selection tools are generally available. Mechanistic network models are algorithmic descriptions of network formation, and they are defined by a small number of domain-specific rules that are informed by our scientific understanding of the problem. Their likelihood functions are generally not analytically tractable, and inference and model selection tools have traditionally not been developed for them. Well-known examples of this model class include the Price model (Price, 1965), the Barabasi-Albert model (Barabasi and Albert, 1999), the Watts-Strogatz model (Watts and Strogatz, 1998), and many others (Sole et al, 2002; Vazquez et al, 2003; Klemm and Eguiluz, 2002; Kumpula et al, 2007)

Objectives
Methods
Findings
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call