Abstract

In large systems where the physical laws governing system behavior are inherently probabilistic and complicated, traditional methods of obtaining closed-form analytic solutions may not be adequate for the level of detailed study needed. In such situations, one could simulate the behavior of the system in order to estimate the desired solution. This approach of using randomized simulations on computers, which came to be called the Monte Carlo methods, is powerful, elegant, flexible, and easy to implement. From the early days of their application in simulating neutron diffusion evolution in fissionable material in the late 1940s, Monte Carlo methods have found application in almost every discipline of science and engineering. Central to the Monte Carlo approach is the generation of a series of random numbers, often a sequence of numbers between 0 and 1 sampled from a uniform distribution, or, in several other instances, a sequence of random numbers sampled from other standard distributions (e.g., the normal distribution) or from more general probability distributions that arise in physical models. More sophisticated techniques are needed for sampling from more general distributions, one such technique being acceptance–rejection sampling. These techniques, however, are not well suited for sampling large-dimensional probability distributions. This situation can be alleviated through the use of Markov chains, in which case the approach is referred to as the Markov chain Monte Carlo (MCMC) method [1]. Typically, MCMC methods refer to a collection of related algorithms, namely, the Metropolis–Hastings algorithm, simulated annealing, and Gibbs sampling.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call