Monte Carlo Method

Download as pdf or txt
Download as pdf or txt
You are on page 1of 23

Monte Carlo method

Monte Carlo methods, or Monte Carlo experiments, are a broad class of computational algorithms that
rely on repeated random sampling to obtain numerical results. The underlying concept is to use randomness
to solve problems that might be deterministic in principle. They are often used in physical and mathematical
problems and are most useful when it is difficult or impossible to use other approaches. Monte Carlo
methods are mainly used in three problem classes:[1] optimization, numerical integration, and generating
draws from a probability distribution.

In physics-related problems, Monte Carlo methods are useful for simulating systems with many coupled
degrees of freedom, such as fluids, disordered materials, strongly coupled solids, and cellular structures (see
cellular Potts model, interacting particle systems, McKean–Vlasov processes, kinetic models of gases).

Other examples include modeling phenomena with significant uncertainty in inputs such as the calculation
of risk in business and, in mathematics, evaluation of multidimensional definite integrals with complicated
boundary conditions. In application to systems engineering problems (space, oil exploration, aircraft design,
etc.), Monte Carlo–based predictions of failure, cost overruns and schedule overruns are routinely better
than human intuition or alternative "soft" methods.[2]

In principle, Monte Carlo methods can be used to solve any problem having a probabilistic interpretation.
By the law of large numbers, integrals described by the expected value of some random variable can be
approximated by taking the empirical mean (a.k.a. the 'sample mean') of independent samples of the
variable. When the probability distribution of the variable is parameterized, mathematicians often use a
Markov chain Monte Carlo (MCMC) sampler.[3][4][5] The central idea is to design a judicious Markov
chain model with a prescribed stationary probability distribution. That is, in the limit, the samples being
generated by the MCMC method will be samples from the desired (target) distribution.[6][7] By the ergodic
theorem, the stationary distribution is approximated by the empirical measures of the random states of the
MCMC sampler.

In other problems, the objective is generating draws from a sequence of probability distributions satisfying a
nonlinear evolution equation. These flows of probability distributions can always be interpreted as the
distributions of the random states of a Markov process whose transition probabilities depend on the
distributions of the current random states (see McKean–Vlasov processes, nonlinear filtering equation).[8][9]
In other instances we are given a flow of probability distributions with an increasing level of sampling
complexity (path spaces models with an increasing time horizon, Boltzmann–Gibbs measures associated
with decreasing temperature parameters, and many others). These models can also be seen as the evolution
of the law of the random states of a nonlinear Markov chain.[9][10] A natural way to simulate these
sophisticated nonlinear Markov processes is to sample multiple copies of the process, replacing in the
evolution equation the unknown distributions of the random states by the sampled empirical measures. In
contrast with traditional Monte Carlo and MCMC methodologies, these mean-field particle techniques rely
on sequential interacting samples. The terminology mean field reflects the fact that each of the samples
(a.k.a. particles, individuals, walkers, agents, creatures, or phenotypes) interacts with the empirical
measures of the process. When the size of the system tends to infinity, these random empirical measures
converge to the deterministic distribution of the random states of the nonlinear Markov chain, so that the
statistical interaction between particles vanishes.

Despite its conceptual and algorithmic simplicity, the computational cost associated with a Monte Carlo
simulation can be staggeringly high. In general the method requires many samples to get a good
approximation, which may incur an arbitrarily large total runtime if the processing time of a single sample is
high.[11] Although this is a severe limitation in very complex problems, the embarrassingly parallel nature
of the algorithm allows this large cost to be reduced (perhaps to a feasible level) through parallel computing
strategies in local processors, clusters, cloud computing, GPU, FPGA, etc.[12][13][14][15]

Overview
Monte Carlo methods vary, but tend to follow a particular pattern:

1. Define a domain of possible inputs


2. Generate inputs randomly from a probability distribution over the domain
3. Perform a deterministic computation on the inputs
4. Aggregate the results

For example, consider a quadrant (circular sector) inscribed in a unit


square. Given that the ratio of their areas is π4 , the value of π can be
approximated using a Monte Carlo method:[16]

1. Draw a square, then inscribe a quadrant within it


2. Uniformly scatter a given number of points over the
square
3. Count the number of points inside the quadrant, i.e.
having a distance from the origin of less than 1
4. The ratio of the inside-count and the total-sample-count
π
is an estimate of the ratio of the two areas, 4 . Multiply the Monte Carlo method applied to
result by 4 to estimate π . approximating the value of π.

In this procedure the domain of inputs is the square that


circumscribes the quadrant. We generate random inputs by scattering grains over the square then perform a
computation on each input (test whether it falls within the quadrant). Aggregating the results yields our final
result, the approximation of π .

There are two important considerations:

1. If the points are not uniformly distributed, then the approximation will be poor.
2. The approximation is generally poor if only a few points are randomly placed in the whole
square. On average, the approximation improves as more points are placed.

Uses of Monte Carlo methods require large amounts of random numbers, and their use benefitted greatly
from pseudorandom number generators, which were far quicker to use than the tables of random numbers
that had been previously used for statistical sampling.

History
Before the Monte Carlo method was developed, simulations tested a previously understood deterministic
problem, and statistical sampling was used to estimate uncertainties in the simulations. Monte Carlo
simulations invert this approach, solving deterministic problems using probabilistic metaheuristics (see
simulated annealing).
An early variant of the Monte Carlo method was devised to solve the Buffon's needle problem, in which π
can be estimated by dropping needles on a floor made of parallel equidistant strips. In the 1930s, Enrico
Fermi first experimented with the Monte Carlo method while studying neutron diffusion, but he did not
publish this work.[17]

In the late 1940s, Stanislaw Ulam invented the modern version of the Markov Chain Monte Carlo method
while he was working on nuclear weapons projects at the Los Alamos National Laboratory. In 1946,
nuclear weapons physicists at Los Alamos were investigating neutron diffusion in the core of a nuclear
weapon.[17] Despite having most of the necessary data, such as the average distance a neutron would travel
in a substance before it collided with an atomic nucleus and how much energy the neutron was likely to
give off following a collision, the Los Alamos physicists were unable to solve the problem using
conventional, deterministic mathematical methods. Ulam proposed using random experiments. He recounts
his inspiration as follows:

The first thoughts and attempts I made to practice [the Monte Carlo Method] were suggested
by a question which occurred to me in 1946 as I was convalescing from an illness and playing
solitaires. The question was what are the chances that a Canfield solitaire laid out with 52
cards will come out successfully? After spending a lot of time trying to estimate them by pure
combinatorial calculations, I wondered whether a more practical method than "abstract
thinking" might not be to lay it out say one hundred times and simply observe and count the
number of successful plays. This was already possible to envisage with the beginning of the
new era of fast computers, and I immediately thought of problems of neutron diffusion and
other questions of mathematical physics, and more generally how to change processes
described by certain differential equations into an equivalent form interpretable as a succession
of random operations. Later [in 1946], I described the idea to John von Neumann, and we
began to plan actual calculations.[18]

Being secret, the work of von Neumann and Ulam required a code name.[19] A colleague of von Neumann
and Ulam, Nicholas Metropolis, suggested using the name Monte Carlo, which refers to the Monte Carlo
Casino in Monaco where Ulam's uncle would borrow money from relatives to gamble.[17] Monte Carlo
methods were central to the simulations required for the Manhattan Project, though severely limited by the
computational tools at the time. Von Neumann, Nicholas Metropolis and others programmed the ENIAC
computer to perform the first fully automated Monte Carlo calculations, of a fission weapon core, in the
spring of 1948.[20] In the 1950s Monte Carlo methods were used at Los Alamos for the development of the
hydrogen bomb, and became popularized in the fields of physics, physical chemistry, and operations
research. The Rand Corporation and the U.S. Air Force were two of the major organizations responsible
for funding and disseminating information on Monte Carlo methods during this time, and they began to find
a wide application in many different fields.

The theory of more sophisticated mean-field type particle Monte Carlo methods had certainly started by the
mid-1960s, with the work of Henry P. McKean Jr. on Markov interpretations of a class of nonlinear
parabolic partial differential equations arising in fluid mechanics.[21][22] We also quote an earlier pioneering
article by Theodore E. Harris and Herman Kahn, published in 1951, using mean-field genetic-type Monte
Carlo methods for estimating particle transmission energies.[23] Mean-field genetic type Monte Carlo
methodologies are also used as heuristic natural search algorithms (a.k.a. metaheuristic) in evolutionary
computing. The origins of these mean-field computational techniques can be traced to 1950 and 1954 with
the work of Alan Turing on genetic type mutation-selection learning machines[24] and the articles by Nils
Aall Barricelli at the Institute for Advanced Study in Princeton, New Jersey.[25][26]
Quantum Monte Carlo, and more specifically diffusion Monte Carlo methods can also be interpreted as a
mean-field particle Monte Carlo approximation of Feynman–Kac path integrals.[27][28][29][30][31][32][33]
The origins of Quantum Monte Carlo methods are often attributed to Enrico Fermi and Robert Richtmyer
who developed in 1948 a mean-field particle interpretation of neutron-chain reactions,[34] but the first
heuristic-like and genetic type particle algorithm (a.k.a. Resampled or Reconfiguration Monte Carlo
methods) for estimating ground state energies of quantum systems (in reduced matrix models) is due to Jack
H. Hetherington in 1984[33] In molecular chemistry, the use of genetic heuristic-like particle methodologies
(a.k.a. pruning and enrichment strategies) can be traced back to 1955 with the seminal work of Marshall N.
Rosenbluth and Arianna W. Rosenbluth.[35]

The use of Sequential Monte Carlo in advanced signal processing and Bayesian inference is more recent. It
was in 1993, that Gordon et al., published in their seminal work[36] the first application of a Monte Carlo
resampling algorithm in Bayesian statistical inference. The authors named their algorithm 'the bootstrap
filter', and demonstrated that compared to other filtering methods, their bootstrap algorithm does not require
any assumption about that state-space or the noise of the system. We also quote another pioneering article in
this field of Genshiro Kitagawa on a related "Monte Carlo filter",[37] and the ones by Pierre Del Moral[38]
and Himilcon Carvalho, Pierre Del Moral, André Monin and Gérard Salut[39] on particle filters published
in the mid-1990s. Particle filters were also developed in signal processing in 1989–1992 by P. Del Moral, J.
C. Noyer, G. Rigal, and G. Salut in the LAAS-CNRS in a series of restricted and classified research reports
with STCAN (Service Technique des Constructions et Armes Navales), the IT company DIGILOG, and
the LAAS-CNRS (https://www.laas.fr/public/en) (the Laboratory for Analysis and Architecture of Systems)
on radar/sonar and GPS signal processing problems.[40][41][42][43][44][45] These Sequential Monte Carlo
methodologies can be interpreted as an acceptance-rejection sampler equipped with an interacting recycling
mechanism.

From 1950 to 1996, all the publications on Sequential Monte Carlo methodologies, including the pruning
and resample Monte Carlo methods introduced in computational physics and molecular chemistry, present
natural and heuristic-like algorithms applied to different situations without a single proof of their
consistency, nor a discussion on the bias of the estimates and on genealogical and ancestral tree based
algorithms. The mathematical foundations and the first rigorous analysis of these particle algorithms were
written by Pierre Del Moral in 1996.[38][46]

Branching type particle methodologies with varying population sizes were also developed in the end of the
1990s by Dan Crisan, Jessica Gaines and Terry Lyons,[47][48][49] and by Dan Crisan, Pierre Del Moral and
Terry Lyons.[50] Further developments in this field were described in 1999 to 2001 by P. Del Moral, A.
Guionnet and L. Miclo.[28][51][52]

Definitions
There is no consensus on how Monte Carlo should be defined. For example, Ripley[53] defines most
probabilistic modeling as stochastic simulation, with Monte Carlo being reserved for Monte Carlo
integration and Monte Carlo statistical tests. Sawilowsky[54] distinguishes between a simulation, a Monte
Carlo method, and a Monte Carlo simulation: a simulation is a fictitious representation of reality, a Monte
Carlo method is a technique that can be used to solve a mathematical or statistical problem, and a Monte
Carlo simulation uses repeated sampling to obtain the statistical properties of some phenomenon (or
behavior). Examples:

Simulation: Drawing one pseudo-random uniform variable from the interval [0,1] can be used
to simulate the tossing of a coin: If the value is less than or equal to 0.50 designate the
outcome as heads, but if the value is greater than 0.50 designate the outcome as tails. This
is a simulation, but not a Monte Carlo simulation.
Monte Carlo method: Pouring out a box of coins on a table, and then computing the ratio of
coins that land heads versus tails is a Monte Carlo method of determining the behavior of
repeated coin tosses, but it is not a simulation.
Monte Carlo simulation: Drawing a large number of pseudo-random uniform variables from
the interval [0,1] at one time, or once at many different times, and assigning values less than
or equal to 0.50 as heads and greater than 0.50 as tails, is a Monte Carlo simulation of the
behavior of repeatedly tossing a coin.

Kalos and Whitlock[55] point out that such distinctions are not always easy to maintain. For example, the
emission of radiation from atoms is a natural stochastic process. It can be simulated directly, or its average
behavior can be described by stochastic equations that can themselves be solved using Monte Carlo
methods. "Indeed, the same computer code can be viewed simultaneously as a 'natural simulation' or as a
solution of the equations by natural sampling."

Monte Carlo and random numbers

The main idea behind this method is that the results are computed based on repeated random sampling and
statistical analysis. The Monte Carlo simulation is, in fact, random experimentations, in the case that, the
results of these experiments are not well known. Monte Carlo simulations are typically characterized by
many unknown parameters, many of which are difficult to obtain experimentally.[56] Monte Carlo
simulation methods do not always require truly random numbers to be useful (although, for some
applications such as primality testing, unpredictability is vital).[57] Many of the most useful techniques use
deterministic, pseudorandom sequences, making it easy to test and re-run simulations. The only quality
usually necessary to make good simulations is for the pseudo-random sequence to appear "random enough"
in a certain sense.

What this means depends on the application, but typically they should pass a series of statistical tests.
Testing that the numbers are uniformly distributed or follow another desired distribution when a large
enough number of elements of the sequence are considered is one of the simplest and most common ones.
Weak correlations between successive samples are also often desirable/necessary.

Sawilowsky lists the characteristics of a high-quality Monte Carlo simulation:[54]

the (pseudo-random) number generator has certain characteristics (e.g. a long "period"
before the sequence repeats)
the (pseudo-random) number generator produces values that pass tests for randomness
there are enough samples to ensure accurate results
the proper sampling technique is used
the algorithm used is valid for what is being modeled
it simulates the phenomenon in question.

Pseudo-random number sampling algorithms are used to transform uniformly distributed pseudo-random
numbers into numbers that are distributed according to a given probability distribution.

Low-discrepancy sequences are often used instead of random sampling from a space as they ensure even
coverage and normally have a faster order of convergence than Monte Carlo simulations using random or
pseudorandom sequences. Methods based on their use are called quasi-Monte Carlo methods.

In an effort to assess the impact of random number quality on Monte Carlo simulation outcomes,
astrophysical researchers tested cryptographically-secure pseudorandom numbers generated via Intel's
RDRAND instruction set, as compared to those derived from algorithms, like the Mersenne Twister, in
Monte Carlo simulations of radio flares from brown dwarfs. RDRAND is the closest pseudorandom
number generator to a true random number generator. No statistically significant difference was found
between models generated with typical pseudorandom number generators and RDRAND for trials
consisting of the generation of 107 random numbers.[58]

Monte Carlo simulation versus "what if" scenarios

There are ways of using probabilities that are definitely not Monte Carlo simulations – for example,
deterministic modeling using single-point estimates. Each uncertain variable within a model is assigned a
"best guess" estimate. Scenarios (such as best, worst, or most likely case) for each input variable are chosen
and the results recorded.[59]

By contrast, Monte Carlo simulations sample from a probability distribution for each variable to produce
hundreds or thousands of possible outcomes. The results are analyzed to get probabilities of different
outcomes occurring.[60] For example, a comparison of a spreadsheet cost construction model run using
traditional "what if" scenarios, and then running the comparison again with Monte Carlo simulation and
triangular probability distributions shows that the Monte Carlo analysis has a narrower range than the "what
if" analysis. This is because the "what if" analysis gives equal weight to all scenarios (see quantifying
uncertainty in corporate finance), while the Monte Carlo method hardly samples in the very low probability
regions. The samples in such regions are called "rare events".

Convergence
Convergence can be checked with the Gelman-Rubin diagnostic or the Geweke diagnostic.

Applications
Monte Carlo methods are especially useful for simulating phenomena with significant uncertainty in inputs
and systems with many coupled degrees of freedom. Areas of application include:

Physical sciences

Monte Carlo methods are very important in computational physics, physical chemistry, and related applied
fields, and have diverse applications from complicated quantum chromodynamics calculations to designing
heat shields and aerodynamic forms as well as in modeling radiation transport for radiation dosimetry
calculations.[61][62][63] In statistical physics, Monte Carlo molecular modeling is an alternative to
computational molecular dynamics, and Monte Carlo methods are used to compute statistical field theories
of simple particle and polymer systems.[35][64] Quantum Monte Carlo methods solve the many-body
problem for quantum systems.[8][9][27] In radiation materials science, the binary collision approximation for
simulating ion implantation is usually based on a Monte Carlo approach to select the next colliding
atom.[65] In experimental particle physics, Monte Carlo methods are used for designing detectors,
understanding their behavior and comparing experimental data to theory. In astrophysics, they are used in
such diverse manners as to model both galaxy evolution[66] and microwave radiation transmission through
a rough planetary surface.[67] Monte Carlo methods are also used in the ensemble models that form the
basis of modern weather forecasting.

Engineering
Monte Carlo methods are widely used in engineering for sensitivity analysis and quantitative probabilistic
analysis in process design. The need arises from the interactive, co-linear and non-linear behavior of typical
process simulations. For example,

In microelectronics engineering, Monte Carlo methods are applied to analyze correlated and
uncorrelated variations in analog and digital integrated circuits.
In geostatistics and geometallurgy, Monte Carlo methods underpin the design of mineral
processing flowsheets and contribute to quantitative risk analysis.[19]
In fluid dynamics, in particular rarefied gas dynamics, where the Boltzmann equation is
solved for finite Knudsen number fluid flows using the direct simulation Monte Carlo[68]
method in combination with highly efficient computational algorithms.[69]
In autonomous robotics, Monte Carlo localization can determine the position of a robot. It is
often applied to stochastic filters such as the Kalman filter or particle filter that forms the heart
of the SLAM (simultaneous localization and mapping) algorithm.
In telecommunications, when planning a wireless network, the design must be proven to
work for a wide variety of scenarios that depend mainly on the number of users, their
locations and the services they want to use. Monte Carlo methods are typically used to
generate these users and their states. The network performance is then evaluated and, if
results are not satisfactory, the network design goes through an optimization process.
In reliability engineering, Monte Carlo simulation is used to compute system-level response
given the component-level response.
In signal processing and Bayesian inference, particle filters and sequential Monte Carlo
techniques are a class of mean-field particle methods for sampling and computing the
posterior distribution of a signal process given some noisy and partial observations using
interacting empirical measures.[70]

Climate change and radiative forcing

The Intergovernmental Panel on Climate Change relies on Monte Carlo methods in probability density
function analysis of radiative forcing.[71]

Computational biology

Monte Carlo methods are used in various fields of computational biology, for example for Bayesian
inference in phylogeny, or for studying biological systems such as genomes, proteins,[72] or membranes.[73]
The systems can be studied in the coarse-grained or ab initio frameworks depending on the desired
accuracy. Computer simulations allow us to monitor the local environment of a particular molecule to see if
some chemical reaction is happening for instance. In cases where it is not feasible to conduct a physical
experiment, thought experiments can be conducted (for instance: breaking bonds, introducing impurities at
specific sites, changing the local/global structure, or introducing external fields).

Computer graphics
Path tracing, occasionally referred to as Monte Carlo ray tracing, renders a 3D scene by randomly tracing
samples of possible light paths. Repeated sampling of any given pixel will eventually cause the average of
the samples to converge on the correct solution of the rendering equation, making it one of the most
physically accurate 3D graphics rendering methods in existence.

Applied statistics

The standards for Monte Carlo experiments in statistics were set by Sawilowsky.[74] In applied statistics,
Monte Carlo methods may be used for at least four purposes:

1. To compare competing statistics for small samples under realistic data conditions. Although
type I error and power properties of statistics can be calculated for data drawn from classical
theoretical distributions (e.g., normal curve, Cauchy distribution) for asymptotic conditions (i.
e, infinite sample size and infinitesimally small treatment effect), real data often do not have
such distributions.[75]
2. To provide implementations of hypothesis tests that are more efficient than exact tests such
as permutation tests (which are often impossible to compute) while being more accurate
than critical values for asymptotic distributions.
3. To provide a random sample from the posterior distribution in Bayesian inference. This
sample then approximates and summarizes all the essential features of the posterior.
4. To provide efficient random estimates of the Hessian matrix of the negative log-likelihood
function that may be averaged to form an estimate of the Fisher information matrix.[76][77]

Monte Carlo methods are also a compromise between approximate randomization and permutation tests.
An approximate randomization test is based on a specified subset of all permutations (which entails
potentially enormous housekeeping of which permutations have been considered). The Monte Carlo
approach is based on a specified number of randomly drawn permutations (exchanging a minor loss in
precision if a permutation is drawn twice—or more frequently—for the efficiency of not having to track
which permutations have already been selected).

Artificial intelligence for games

Monte Carlo methods have been developed into a technique called Monte-Carlo tree search that is useful
for searching for the best move in a game. Possible moves are organized in a search tree and many random
simulations are used to estimate the long-term potential of each move. A black box simulator represents the
opponent's moves.[78]

The Monte Carlo tree search (MCTS) method has four steps:[79]

1. Starting at root node of the tree, select optimal child nodes until a leaf node is reached.
2. Expand the leaf node and choose one of its children.
3. Play a simulated game starting with that node.
4. Use the results of that simulated game to update the node and its ancestors.

The net effect, over the course of many simulated games, is that the value of a node representing a move
will go up or down, hopefully corresponding to whether or not that node represents a good move.
Monte Carlo Tree Search has been used successfully to play games such as Go,[80] Tantrix,[81]
Battleship,[82] Havannah,[83] and Arimaa.[84]

Design and visuals

Monte Carlo methods are also efficient in solving coupled integral differential equations of radiation fields
and energy transport, and thus these methods have been used in global illumination computations that
produce photo-realistic images of virtual 3D models, with applications in video games, architecture, design,
computer generated films, and cinematic special effects.[85]

Search and rescue

The US Coast Guard utilizes Monte Carlo methods within its computer modeling software SAROPS in
order to calculate the probable locations of vessels during search and rescue operations. Each simulation
can generate as many as ten thousand data points that are randomly distributed based upon provided
variables.[86] Search patterns are then generated based upon extrapolations of these data in order to
optimize the probability of containment (POC) and the probability of detection (POD), which together will
equal an overall probability of success (POS). Ultimately this serves as a practical application of probability
distribution in order to provide the swiftest and most expedient method of rescue, saving both lives and
resources.[87]

Finance and business

Monte Carlo simulation is commonly used to evaluate the risk and uncertainty that would affect the
outcome of different decision options. Monte Carlo simulation allows the business risk analyst to
incorporate the total effects of uncertainty in variables like sales volume, commodity and labour prices,
interest and exchange rates, as well as the effect of distinct risk events like the cancellation of a contract or
the change of a tax law.

Monte Carlo methods in finance are often used to evaluate investments in projects at a business unit or
corporate level, or other financial valuations. They can be used to model project schedules, where
simulations aggregate estimates for worst-case, best-case, and most likely durations for each task to
determine outcomes for the overall project.[1] (https://risk.octigo.pl/) Monte Carlo methods are also used in
option pricing, default risk analysis.[88][89][90] Additionally, they can be used to estimate the financial
impact of medical interventions.[91]

Law

A Monte Carlo approach was used for evaluating the potential value of a proposed program to help female
petitioners in Wisconsin be successful in their applications for harassment and domestic abuse restraining
orders. It was proposed to help women succeed in their petitions by providing them with greater advocacy
thereby potentially reducing the risk of rape and physical assault. However, there were many variables in
play that could not be estimated perfectly, including the effectiveness of restraining orders, the success rate
of petitioners both with and without advocacy, and many others. The study ran trials that varied these
variables to come up with an overall estimate of the success level of the proposed program as a whole.[92]

Library science

Monte Carlo approach had also been used to simulate the number of book publications based on book
genre in Malaysia. The Monte Carlo simulation utilized previous published National Book publication data
and book's price according to book genre in the local market. The Monte Carlo results were used to
determine what kind of book genre that Malaysians are fond of and was used to compare book publications
between Malaysia and Japan.[93]

Other

Nassim Nicholas Taleb writes about Monte Carlo generators in his 2001 book Fooled by Randomness as a
real instance of the reverse Turing test: a human can be declared unintelligent if their writing cannot be told
apart from a generated one.

Use in mathematics
In general, the Monte Carlo methods are used in mathematics to solve various problems by generating
suitable random numbers (see also Random number generation) and observing that fraction of the numbers
that obeys some property or properties. The method is useful for obtaining numerical solutions to problems
too complicated to solve analytically. The most common application of the Monte Carlo method is Monte
Carlo integration.

Integration

Deterministic numerical integration algorithms work well in a small


number of dimensions, but encounter two problems when the
functions have many variables. First, the number of function
evaluations needed increases rapidly with the number of
dimensions. For example, if 10 evaluations provide adequate
accuracy in one dimension, then 10100 points are needed for 100
dimensions—far too many to be computed. This is called the curse
of dimensionality. Second, the boundary of a multidimensional
region may be very complicated, so it may not be feasible to reduce
the problem to an iterated integral.[94] 100 dimensions is by no
means unusual, since in many physical problems, a "dimension" is
equivalent to a degree of freedom. Monte-Carlo integration works by
comparing random points with the
Monte Carlo methods provide a way out of this exponential value of the function
increase in computation time. As long as the function in question is
reasonably well-behaved, it can be estimated by randomly selecting
points in 100-dimensional space, and taking some kind of average of the function values at these points. By
the central limit theorem, this method displays convergence—i.e., quadrupling the number of sampled
points halves the error, regardless of the number of dimensions.[94]
A refinement of this method, known as importance sampling in
statistics, involves sampling the points randomly, but more
frequently where the integrand is large. To do this precisely one
would have to already know the integral, but one can approximate
the integral by an integral of a similar function or use adaptive
routines such as stratified sampling, recursive stratified sampling,
adaptive umbrella sampling[95][96] or the VEGAS algorithm.

A similar approach, the quasi-Monte Carlo method, uses low-


discrepancy sequences. These sequences "fill" the area better and
sample the most important points more frequently, so quasi-Monte Errors reduce by a factor of
Carlo methods can often converge on the integral more quickly.

Another class of methods for sampling points in a volume is to simulate random walks over it (Markov
chain Monte Carlo). Such methods include the Metropolis–Hastings algorithm, Gibbs sampling, Wang and
Landau algorithm, and interacting type MCMC methodologies such as the sequential Monte Carlo
samplers.[97]

Simulation and optimization

Another powerful and very popular application for random numbers in numerical simulation is in numerical
optimization. The problem is to minimize (or maximize) functions of some vector that often has many
dimensions. Many problems can be phrased in this way: for example, a computer chess program could be
seen as trying to find the set of, say, 10 moves that produces the best evaluation function at the end. In the
traveling salesman problem the goal is to minimize distance traveled. There are also applications to
engineering design, such as multidisciplinary design optimization. It has been applied with quasi-one-
dimensional models to solve particle dynamics problems by efficiently exploring large configuration space.
Reference[98] is a comprehensive review of many issues related to simulation and optimization.

The traveling salesman problem is what is called a conventional optimization problem. That is, all the facts
(distances between each destination point) needed to determine the optimal path to follow are known with
certainty and the goal is to run through the possible travel choices to come up with the one with the lowest
total distance. However, let's assume that instead of wanting to minimize the total distance traveled to visit
each desired destination, we wanted to minimize the total time needed to reach each destination. This goes
beyond conventional optimization since travel time is inherently uncertain (traffic jams, time of day, etc.).
As a result, to determine our optimal path we would want to use simulation - optimization to first
understand the range of potential times it could take to go from one point to another (represented by a
probability distribution in this case rather than a specific distance) and then optimize our travel decisions to
identify the best path to follow taking that uncertainty into account.

Inverse problems

Probabilistic formulation of inverse problems leads to the definition of a probability distribution in the
model space. This probability distribution combines prior information with new information obtained by
measuring some observable parameters (data). As, in the general case, the theory linking data with model
parameters is nonlinear, the posterior probability in the model space may not be easy to describe (it may be
multimodal, some moments may not be defined, etc.).

When analyzing an inverse problem, obtaining a maximum likelihood model is usually not sufficient, as we
normally also wish to have information on the resolution power of the data. In the general case we may
have many model parameters, and an inspection of the marginal probability densities of interest may be
impractical, or even useless. But it is possible to pseudorandomly generate a large collection of models
according to the posterior probability distribution and to analyze and display the models in such a way that
information on the relative likelihoods of model properties is conveyed to the spectator. This can be
accomplished by means of an efficient Monte Carlo method, even in cases where no explicit formula for
the a priori distribution is available.

The best-known importance sampling method, the Metropolis algorithm, can be generalized, and this gives
a method that allows analysis of (possibly highly nonlinear) inverse problems with complex a priori
information and data with an arbitrary noise distribution.[99][100]

Philosophy

Popular exposition of the Monte Carlo Method was conducted by McCracken.[101] The method's general
philosophy was discussed by Elishakoff[102] and Grüne-Yanoff and Weirich.[103]

See also
Mathematics
portal

Auxiliary field Monte Carlo


Biology Monte Carlo method
Direct simulation Monte Carlo
Dynamic Monte Carlo method
Ergodicity
Genetic algorithms
Kinetic Monte Carlo
List of software for Monte Carlo molecular modeling
Mean-field particle methods
Monte Carlo method for photon transport
Monte Carlo methods for electron transport
Monte Carlo N-Particle Transport Code
Morris method
Multilevel Monte Carlo method
Quasi-Monte Carlo method
Sobol sequence
Temporal difference learning

References

Citations
1. Kroese, D. P.; Brereton, T.; Taimre, T.; Botev, Z. I. (2014). "Why the Monte Carlo method is so
important today" (https://semanticscholar.org/paper/7a56b632de84d0b81f283750b11609a0
42890639). WIREs Comput Stat. 6 (6): 386–392. doi:10.1002/wics.1314 (https://doi.org/10.1
002%2Fwics.1314). S2CID 18521840 (https://api.semanticscholar.org/CorpusID:18521840).
2. Hubbard, Douglas; Samuelson, Douglas A. (October 2009). "Modeling Without
Measurements" (http://viewer.zmags.com/publication/357348e6#/357348e6/28). OR/MS
Today: 28–33.
3. Metropolis, Nicholas; Rosenbluth, Arianna W.; Rosenbluth, Marshall N.; Teller, Augusta H.;
Teller, Edward (1 June 1953). "Equation of State Calculations by Fast Computing Machines"
(https://semanticscholar.org/paper/f6a13f116e270dde9d67848495f801cdb8efa25d). The
Journal of Chemical Physics. 21 (6): 1087–1092. Bibcode:1953JChPh..21.1087M (https://ui.
adsabs.harvard.edu/abs/1953JChPh..21.1087M). doi:10.1063/1.1699114 (https://doi.org/10.
1063%2F1.1699114). ISSN 0021-9606 (https://www.worldcat.org/issn/0021-9606).
OSTI 4390578 (https://www.osti.gov/biblio/4390578). S2CID 1046577 (https://api.semantics
cholar.org/CorpusID:1046577).
4. Hastings, W. K. (1 April 1970). "Monte Carlo sampling methods using Markov chains and
their applications" (https://semanticscholar.org/paper/143d2e02ab91ae6259576ac50b664b8
647af8988). Biometrika. 57 (1): 97–109. Bibcode:1970Bimka..57...97H (https://ui.adsabs.har
vard.edu/abs/1970Bimka..57...97H). doi:10.1093/biomet/57.1.97 (https://doi.org/10.1093%2F
biomet%2F57.1.97). ISSN 0006-3444 (https://www.worldcat.org/issn/0006-3444).
S2CID 21204149 (https://api.semanticscholar.org/CorpusID:21204149).
5. Liu, Jun S.; Liang, Faming; Wong, Wing Hung (1 March 2000). "The Multiple-Try Method and
Local Optimization in Metropolis Sampling" (https://semanticscholar.org/paper/ff17c129a8d3
2bb7dc7206230da612e94bd24b9f). Journal of the American Statistical Association. 95
(449): 121–134. doi:10.1080/01621459.2000.10473908 (https://doi.org/10.1080%2F016214
59.2000.10473908). ISSN 0162-1459 (https://www.worldcat.org/issn/0162-1459).
S2CID 123468109 (https://api.semanticscholar.org/CorpusID:123468109).
6. Spall, J. C. (2003). "Estimation via Markov Chain Monte Carlo". IEEE Control Systems
Magazine. 23 (2): 34–45. doi:10.1109/MCS.2003.1188770 (https://doi.org/10.1109%2FMCS.
2003.1188770).
7. Hill, Stacy D.; Spall, James C. (2019). "Stationarity and Convergence of the Metropolis-
Hastings Algorithm: Insights into Theoretical Aspects". IEEE Control Systems Magazine. 39:
56–67. doi:10.1109/MCS.2018.2876959 (https://doi.org/10.1109%2FMCS.2018.2876959).
S2CID 58672766 (https://api.semanticscholar.org/CorpusID:58672766).
8. Kolokoltsov, Vassili (2010). Nonlinear Markov processes. Cambridge University Press.
p. 375.
9. Del Moral, Pierre (2013). Mean field simulation for Monte Carlo integration (http://www.crcpre
ss.com/product/isbn/9781466504059). Chapman & Hall/CRC Press. p. 626. "Monographs
on Statistics & Applied Probability"
10. Del Moral, P.; Doucet, A.; Jasra, A. (2006). "Sequential Monte Carlo samplers". Journal of the
Royal Statistical Society, Series B. 68 (3): 411–436. arXiv:cond-mat/0212648 (https://arxiv.or
g/abs/cond-mat/0212648). doi:10.1111/j.1467-9868.2006.00553.x (https://doi.org/10.1111%
2Fj.1467-9868.2006.00553.x). S2CID 12074789 (https://api.semanticscholar.org/CorpusID:1
2074789).
11. Shonkwiler, R. W.; Mendivil, F. (2009). Explorations in Monte Carlo Methods. Springer.
12. Atanassova, E.; Gurov, T.; Karaivanova, A.; Ivanovska, S.; Durchova, M.; Dimitrov, D. (2016).
"On the parallelization approaches for Intel MIC architecture". AIP Conference Proceedings.
1773 (1): 070001. Bibcode:2016AIPC.1773g0001A (https://ui.adsabs.harvard.edu/abs/2016
AIPC.1773g0001A). doi:10.1063/1.4964983 (https://doi.org/10.1063%2F1.4964983).
13. Cunha Jr, A.; Nasser, R.; Sampaio, R.; Lopes, H.; Breitman, K. (2014). "Uncertainty
quantification through the Monte Carlo method in a cloud computing setting". Computer
Physics Communications. 185 (5): 1355–1363. arXiv:2105.09512 (https://arxiv.org/abs/2105.
09512). Bibcode:2014CoPhC.185.1355C (https://ui.adsabs.harvard.edu/abs/2014CoPhC.18
5.1355C). doi:10.1016/j.cpc.2014.01.006 (https://doi.org/10.1016%2Fj.cpc.2014.01.006).
S2CID 32376269 (https://api.semanticscholar.org/CorpusID:32376269).
14. Wei, J.; Kruis, F.E. (2013). "A GPU-based parallelized Monte-Carlo method for particle
coagulation using an acceptance–rejection strategy". Chemical Engineering Science. 104:
451–459. Bibcode:2013ChEnS.104..451W (https://ui.adsabs.harvard.edu/abs/2013ChEnS.1
04..451W). doi:10.1016/j.ces.2013.08.008 (https://doi.org/10.1016%2Fj.ces.2013.08.008).
15. Lin, Y.; Wang, F.; Liu, B. (2018). "Random number generators for large-scale parallel Monte
Carlo simulations on FPGA". Journal of Computational Physics. 360: 93–103.
Bibcode:2018JCoPh.360...93L (https://ui.adsabs.harvard.edu/abs/2018JCoPh.360...93L).
doi:10.1016/j.jcp.2018.01.029 (https://doi.org/10.1016%2Fj.jcp.2018.01.029).
16. Kalos & Whitlock 2008.
17. Metropolis 1987.
18. Eckhardt 1987.
19. Mazhdrakov, Benov & Valkanov 2018, p. 250.
20. Haigh, Thomas; Priestley, Mark; Rope, Crispin (2014). "Los Alamos Bets on ENIAC: Nuclear
Monte Carlo Simulations, 1947-1948" (https://ieeexplore.ieee.org/document/6880250). IEEE
Annals of the History of Computing. 36 (3): 42–63. doi:10.1109/MAHC.2014.40 (https://doi.or
g/10.1109%2FMAHC.2014.40). S2CID 17470931 (https://api.semanticscholar.org/CorpusID:
17470931).
21. McKean, Henry P. (1967). "Propagation of chaos for a class of non-linear parabolic
equations". Lecture Series in Differential Equations, Catholic Univ. 7: 41–57.
22. McKean, Henry P. (1966). "A class of Markov processes associated with nonlinear parabolic
equations" (https://www.ncbi.nlm.nih.gov/pmc/articles/PMC220210). Proc. Natl. Acad. Sci.
USA. 56 (6): 1907–1911. Bibcode:1966PNAS...56.1907M (https://ui.adsabs.harvard.edu/ab
s/1966PNAS...56.1907M). doi:10.1073/pnas.56.6.1907 (https://doi.org/10.1073%2Fpnas.56.
6.1907). PMC 220210 (https://www.ncbi.nlm.nih.gov/pmc/articles/PMC220210).
PMID 16591437 (https://pubmed.ncbi.nlm.nih.gov/16591437).
23. Herman, Kahn; Theodore, Harris E. (1951). "Estimation of particle transmission by random
sampling" (https://dornsifecms.usc.edu/assets/sites/520/docs/kahnharris.pdf) (PDF). Natl.
Bur. Stand. Appl. Math. Ser. 12: 27–30.
24. Turing, Alan M. (1950). "Computing machinery and intelligence". Mind. LIX (238): 433–460.
doi:10.1093/mind/LIX.236.433 (https://doi.org/10.1093%2Fmind%2FLIX.236.433).
25. Barricelli, Nils Aall (1954). "Esempi numerici di processi di evoluzione". Methodos: 45–68.
26. Barricelli, Nils Aall (1957). "Symbiogenetic evolution processes realized by artificial
methods". Methodos: 143–182.
27. Del Moral, Pierre (2004). Feynman–Kac formulae. Genealogical and interacting particle
approximations (https://www.springer.com/mathematics/probability/book/978-0-387-20268-
6). Probability and Its Applications. Springer. p. 575. ISBN 9780387202686. "Series:
Probability and Applications"
28. Del Moral, P.; Miclo, L. (2000). "Branching and interacting particle systems approximations of
Feynman–Kac formulae with applications to non-linear filtering" (http://archive.numdam.org/i
tem/SPS_2000__34__1_0). Séminaire de Probabilités, XXXIV (http://www.numdam.org/ite
m/SPS_2000__34__1_0/). Lecture Notes in Mathematics. Vol. 1729. Berlin: Springer. pp. 1–
145. doi:10.1007/BFb0103798 (https://doi.org/10.1007%2FBFb0103798). ISBN 978-3-540-
67314-9. MR 1768060 (https://mathscinet.ams.org/mathscinet-getitem?mr=1768060).
29. Del Moral, Pierre; Miclo, Laurent (2000). "A Moran particle system approximation of
Feynman–Kac formulae" (https://doi.org/10.1016%2FS0304-4149%2899%2900094-0).
Stochastic Processes and Their Applications. 86 (2): 193–216. doi:10.1016/S0304-
4149(99)00094-0 (https://doi.org/10.1016%2FS0304-4149%2899%2900094-0).
30. Del Moral, Pierre (2003). "Particle approximations of Lyapunov exponents connected to
Schrödinger operators and Feynman–Kac semigroups" (http://journals.cambridge.org/downl
oad.php?file=%2FPSS%2FPSS7%2FS1292810003000016a.pdf&code=a0dbaa7ffca87112
6dc05fe2f918880a). ESAIM Probability & Statistics. 7: 171–208. doi:10.1051/ps:2003001 (ht
tps://doi.org/10.1051%2Fps%3A2003001).
31. Assaraf, Roland; Caffarel, Michel; Khelif, Anatole (2000). "Diffusion Monte Carlo Methods
with a fixed number of walkers" (https://web.archive.org/web/20141107015724/http://qmcche
m.ups-tlse.fr/files/caffarel/31.pdf) (PDF). Phys. Rev. E. 61 (4): 4566–4575.
Bibcode:2000PhRvE..61.4566A (https://ui.adsabs.harvard.edu/abs/2000PhRvE..61.4566A).
doi:10.1103/physreve.61.4566 (https://doi.org/10.1103%2Fphysreve.61.4566).
PMID 11088257 (https://pubmed.ncbi.nlm.nih.gov/11088257). Archived from the original (htt
p://qmcchem.ups-tlse.fr/files/caffarel/31.pdf) (PDF) on 7 November 2014.
32. Caffarel, Michel; Ceperley, David; Kalos, Malvin (1993). "Comment on Feynman–Kac Path-
Integral Calculation of the Ground-State Energies of Atoms". Phys. Rev. Lett. 71 (13): 2159.
Bibcode:1993PhRvL..71.2159C (https://ui.adsabs.harvard.edu/abs/1993PhRvL..71.2159C).
doi:10.1103/physrevlett.71.2159 (https://doi.org/10.1103%2Fphysrevlett.71.2159).
PMID 10054598 (https://pubmed.ncbi.nlm.nih.gov/10054598).
33. Hetherington, Jack H. (1984). "Observations on the statistical iteration of matrices". Phys.
Rev. A. 30 (2713): 2713–2719. Bibcode:1984PhRvA..30.2713H (https://ui.adsabs.harvard.ed
u/abs/1984PhRvA..30.2713H). doi:10.1103/PhysRevA.30.2713 (https://doi.org/10.1103%2F
PhysRevA.30.2713).
34. Fermi, Enrique; Richtmyer, Robert D. (1948). "Note on census-taking in Monte Carlo
calculations" (http://scienze-como.uninsubria.it/bressanini/montecarlo-history/fermi-1948.pd
f) (PDF). LAM. 805 (A). "Declassified report Los Alamos Archive"
35. Rosenbluth, Marshall N.; Rosenbluth, Arianna W. (1955). "Monte-Carlo calculations of the
average extension of macromolecular chains" (https://semanticscholar.org/paper/1570c85ba
9aca1cb413ada31e215e0917c3ccba7). J. Chem. Phys. 23 (2): 356–359.
Bibcode:1955JChPh..23..356R (https://ui.adsabs.harvard.edu/abs/1955JChPh..23..356R).
doi:10.1063/1.1741967 (https://doi.org/10.1063%2F1.1741967). S2CID 89611599 (https://ap
i.semanticscholar.org/CorpusID:89611599).
36. Gordon, N.J.; Salmond, D.J.; Smith, A.F.M. (April 1993). "Novel approach to nonlinear/non-
Gaussian Bayesian state estimation" (https://semanticscholar.org/paper/65484334a5cd4cab
f6e5f7a17f606f07e2acf625). IEE Proceedings F - Radar and Signal Processing. 140 (2):
107–113. doi:10.1049/ip-f-2.1993.0015 (https://doi.org/10.1049%2Fip-f-2.1993.0015).
ISSN 0956-375X (https://www.worldcat.org/issn/0956-375X). S2CID 12644877 (https://api.s
emanticscholar.org/CorpusID:12644877).
37. Kitagawa, G. (1996). "Monte carlo filter and smoother for non-Gaussian nonlinear state
space models". Journal of Computational and Graphical Statistics. 5 (1): 1–25.
doi:10.2307/1390750 (https://doi.org/10.2307%2F1390750). JSTOR 1390750 (https://www.j
stor.org/stable/1390750).
38. Del Moral, Pierre (1996). "Non Linear Filtering: Interacting Particle Solution" (http://web.math
s.unsw.edu.au/~peterdel-moral/mprfs.pdf) (PDF). Markov Processes and Related Fields. 2
(4): 555–580.
39. Carvalho, Himilcon; Del Moral, Pierre; Monin, André; Salut, Gérard (July 1997). "Optimal
Non-linear Filtering in GPS/INS Integration" (http://homepages.laas.fr/monin/Version_anglai
se/Publications_files/GPS.pdf) (PDF). IEEE Transactions on Aerospace and Electronic
Systems. 33 (3): 835–850. Bibcode:1997ITAES..33..835C (https://ui.adsabs.harvard.edu/ab
s/1997ITAES..33..835C). doi:10.1109/7.599254 (https://doi.org/10.1109%2F7.599254).
S2CID 27966240 (https://api.semanticscholar.org/CorpusID:27966240).
40. P. Del Moral, G. Rigal, and G. Salut. "Estimation and nonlinear optimal control: An unified
framework for particle solutions". LAAS-CNRS, Toulouse, Research Report no. 91137,
DRET-DIGILOG- LAAS/CNRS contract, April (1991).
41. P. Del Moral, G. Rigal, and G. Salut. "Nonlinear and non Gaussian particle filters applied to
inertial platform repositioning." LAAS-CNRS, Toulouse, Research Report no. 92207,
STCAN/DIGILOG-LAAS/CNRS Convention STCAN no. A.91.77.013, (94p.) September
(1991).
42. P. Del Moral, G. Rigal, and G. Salut. "Estimation and nonlinear optimal control: Particle
resolution in filtering and estimation: Experimental results". Convention DRET no.
89.34.553.00.470.75.01, Research report no.2 (54p.), January (1992).
43. P. Del Moral, G. Rigal, and G. Salut. "Estimation and nonlinear optimal control: Particle
resolution in filtering and estimation: Theoretical results". Convention DRET no.
89.34.553.00.470.75.01, Research report no.3 (123p.), October (1992).
44. P. Del Moral, J.-Ch. Noyer, G. Rigal, and G. Salut. "Particle filters in radar signal processing:
detection, estimation and air targets recognition". LAAS-CNRS, Toulouse, Research report
no. 92495, December (1992).
45. P. Del Moral, G. Rigal, and G. Salut. "Estimation and nonlinear optimal control: Particle
resolution in filtering and estimation". Studies on: Filtering, optimal control, and maximum
likelihood estimation. Convention DRET no. 89.34.553.00.470.75.01. Research report no.4
(210p.), January (1993).
46. Del Moral, Pierre (1998). "Measure Valued Processes and Interacting Particle Systems.
Application to Non Linear Filtering Problems" (http://projecteuclid.org/download/pdf_1/eucli
d.aoap/1028903535). Annals of Applied Probability (Publications du Laboratoire de
Statistique et Probabilités, 96-15 (1996) ed.). 8 (2): 438–495. CiteSeerX 10.1.1.55.5257 (http
s://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.55.5257).
doi:10.1214/aoap/1028903535 (https://doi.org/10.1214%2Faoap%2F1028903535).
47. Crisan, Dan; Gaines, Jessica; Lyons, Terry (1998). "Convergence of a branching particle
method to the solution of the Zakai" (https://semanticscholar.org/paper/99e8759a243cd0568
b0f32cbace2ad0525b16bb6). SIAM Journal on Applied Mathematics. 58 (5): 1568–1590.
doi:10.1137/s0036139996307371 (https://doi.org/10.1137%2Fs0036139996307371).
S2CID 39982562 (https://api.semanticscholar.org/CorpusID:39982562).
48. Crisan, Dan; Lyons, Terry (1997). "Nonlinear filtering and measure-valued processes" (http
s://doi.org/10.1007%2Fs004400050131). Probability Theory and Related Fields. 109 (2):
217–244. doi:10.1007/s004400050131 (https://doi.org/10.1007%2Fs004400050131).
S2CID 119809371 (https://api.semanticscholar.org/CorpusID:119809371).
49. Crisan, Dan; Lyons, Terry (1999). "A particle approximation of the solution of the Kushner–
Stratonovitch equation" (https://doi.org/10.1007%2Fs004400050249). Probability Theory
and Related Fields. 115 (4): 549–578. doi:10.1007/s004400050249 (https://doi.org/10.100
7%2Fs004400050249). S2CID 117725141 (https://api.semanticscholar.org/CorpusID:11772
5141).
50. Crisan, Dan; Del Moral, Pierre; Lyons, Terry (1999). "Discrete filtering using branching and
interacting particle systems" (http://web.maths.unsw.edu.au/~peterdel-moral/crisan98discret
e.pdf) (PDF). Markov Processes and Related Fields. 5 (3): 293–318.
51. Del Moral, Pierre; Guionnet, Alice (1999). "On the stability of Measure Valued Processes
with Applications to filtering". C. R. Acad. Sci. Paris. 39 (1): 429–434.
52. Del Moral, Pierre; Guionnet, Alice (2001). "On the stability of interacting processes with
applications to filtering and genetic algorithms" (http://web.maths.unsw.edu.au/~peterdel-mor
al/ihp.ps). Annales de l'Institut Henri Poincaré. 37 (2): 155–194.
Bibcode:2001AIHPB..37..155D (https://ui.adsabs.harvard.edu/abs/2001AIHPB..37..155D).
doi:10.1016/s0246-0203(00)01064-5 (https://doi.org/10.1016%2Fs0246-0203%2800%2901
064-5).
53. Ripley 1987
54. Sawilowsky 2003
55. Kalos & Whitlock 2008
56. Shojaeefard, M.H.; Khalkhali, A.; Yarmohammadisatri, Sadegh (2017). "An efficient
sensitivity analysis method for modified geometry of Macpherson suspension based on
Pearson Correlation Coefficient". Vehicle System Dynamics. 55 (6): 827–852.
Bibcode:2017VSD....55..827S (https://ui.adsabs.harvard.edu/abs/2017VSD....55..827S).
doi:10.1080/00423114.2017.1283046 (https://doi.org/10.1080%2F00423114.2017.128304
6). S2CID 114260173 (https://api.semanticscholar.org/CorpusID:114260173).
57. Davenport 1992
58. Route, Matthew (August 10, 2017). "Radio-flaring Ultracool Dwarf Population Synthesis".
The Astrophysical Journal. 845 (1): 66. arXiv:1707.02212 (https://arxiv.org/abs/1707.02212).
Bibcode:2017ApJ...845...66R (https://ui.adsabs.harvard.edu/abs/2017ApJ...845...66R).
doi:10.3847/1538-4357/aa7ede (https://doi.org/10.3847%2F1538-4357%2Faa7ede).
S2CID 118895524 (https://api.semanticscholar.org/CorpusID:118895524).
59. Vose 2008, p. 13.
60. Vose 2008, p. 16.
61. Jia, Xun; Ziegenhein, Peter; Jiang, Steve B (2014). "GPU-based high-performance
computing for radiation therapy" (https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4003902).
Physics in Medicine and Biology. 59 (4): R151–R182. Bibcode:2014PMB....59R.151J (http
s://ui.adsabs.harvard.edu/abs/2014PMB....59R.151J). doi:10.1088/0031-9155/59/4/R151 (htt
ps://doi.org/10.1088%2F0031-9155%2F59%2F4%2FR151). PMC 4003902 (https://www.ncb
i.nlm.nih.gov/pmc/articles/PMC4003902). PMID 24486639 (https://pubmed.ncbi.nlm.nih.gov/
24486639).
62. Hill, R.; Healy, B.; Holloway, L.; Kuncic, Z.; Thwaites, D.; Baldock, C. (Mar 2014). "Advances
in kilovoltage x-ray beam dosimetry" (https://semanticscholar.org/paper/fb231c3d9ade811d7
93b85623fd32c6ea126d5ff). Physics in Medicine and Biology. 59 (6): R183–R231.
Bibcode:2014PMB....59R.183H (https://ui.adsabs.harvard.edu/abs/2014PMB....59R.183H).
doi:10.1088/0031-9155/59/6/R183 (https://doi.org/10.1088%2F0031-9155%2F59%2F6%2F
R183). PMID 24584183 (https://pubmed.ncbi.nlm.nih.gov/24584183). S2CID 18082594 (http
s://api.semanticscholar.org/CorpusID:18082594).
63. Rogers, D.W.O. (2006). "Fifty years of Monte Carlo simulations for medical physics" (https://s
emanticscholar.org/paper/b6d08efc5f0818a01dc60637a4a6f8115482483e). Physics in
Medicine and Biology. 51 (13): R287–R301. Bibcode:2006PMB....51R.287R (https://ui.adsa
bs.harvard.edu/abs/2006PMB....51R.287R). doi:10.1088/0031-9155/51/13/R17 (https://doi.or
g/10.1088%2F0031-9155%2F51%2F13%2FR17). PMID 16790908 (https://pubmed.ncbi.nl
m.nih.gov/16790908). S2CID 12066026 (https://api.semanticscholar.org/CorpusID:1206602
6).
64. Baeurle 2009
65. Möller, W.; Eckstein, W. (1 March 1984). "Tridyn — A TRIM simulation code including
dynamic composition changes". Nuclear Instruments and Methods in Physics Research
Section B: Beam Interactions with Materials and Atoms. 2 (1): 814–818.
Bibcode:1984NIMPB...2..814M (https://ui.adsabs.harvard.edu/abs/1984NIMPB...2..814M).
doi:10.1016/0168-583X(84)90321-5 (https://doi.org/10.1016%2F0168-583X%2884%299032
1-5).
66. MacGillivray & Dodd 1982
67. Golden 1979
68. G. A. Bird, Molecular Gas Dynamics, Clarendon, Oxford (1976)
69. Dietrich, S.; Boyd, I. (1996). "A Scalar optimized parallel implementation of the DSMC
technique" (https://doi.org/10.1006%2Fjcph.1996.0141). Journal of Computational Physics.
126 (2): 328–42. Bibcode:1996JCoPh.126..328D (https://ui.adsabs.harvard.edu/abs/1996JC
oPh.126..328D). doi:10.1006/jcph.1996.0141 (https://doi.org/10.1006%2Fjcph.1996.0141).
70. Chen, Shang-Ying; Hsu, Kuo-Chin; Fan, Chia-Ming (15 March 2021). "Improvement of
generalized finite difference method for stochastic subsurface flow modeling". Journal of
Computational Physics. 429: 110002. Bibcode:2021JCoPh.42910002C (https://ui.adsabs.ha
rvard.edu/abs/2021JCoPh.42910002C). doi:10.1016/J.JCP.2020.110002 (https://doi.org/10.
1016%2FJ.JCP.2020.110002). S2CID 228828681 (https://api.semanticscholar.org/CorpusI
D:228828681).
71. Climate Change 2013 The Physical Science Basis (http://www.climatechange2013.org/imag
es/report/WG1AR5_ALL_FINAL.pdf) (PDF). Cambridge University Press. 2013. p. 697.
ISBN 978-1-107-66182-0. Retrieved 6 July 2023.
72. Ojeda et al. 2009.
73. Milik & Skolnick 1993.
74. Cassey; Smith (2014). "Simulating confidence for the Ellison-Glaeser Index". Journal of
Urban Economics. 81: 93. doi:10.1016/j.jue.2014.02.005 (https://doi.org/10.1016%2Fj.jue.20
14.02.005).
75. Sawilowsky & Fahoome 2003
76. Spall, James C. (2005). "Monte Carlo Computation of the Fisher Information Matrix in
Nonstandard Settings". Journal of Computational and Graphical Statistics. 14 (4): 889–909.
CiteSeerX 10.1.1.142.738 (https://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.142.7
38). doi:10.1198/106186005X78800 (https://doi.org/10.1198%2F106186005X78800).
S2CID 16090098 (https://api.semanticscholar.org/CorpusID:16090098).
77. Das, Sonjoy; Spall, James C.; Ghanem, Roger (2010). "Efficient Monte Carlo computation of
Fisher information matrix using prior information". Computational Statistics & Data Analysis.
54 (2): 272–289. doi:10.1016/j.csda.2009.09.018 (https://doi.org/10.1016%2Fj.csda.2009.09.
018).
78. Chaslot, Guillaume; Bakkes, Sander; Szita, Istvan; Spronck, Pieter. "Monte-Carlo Tree
Search: A New Framework for Game AI" (http://sander.landofsand.com/publications/Monte-C
arlo_Tree_Search_-_A_New_Framework_for_Game_AI.pdf) (PDF).
Sander.landofsand.com. Retrieved 28 October 2017.
79. "Monte Carlo Tree Search - About" (https://web.archive.org/web/20151129023043/http://mct
s.ai/about/index.html). Archived from the original (http://mcts.ai/about/index.html) on 29
November 2015. Retrieved 15 May 2013.
80. Chaslot, Guillaume M. J. -B; Winands, Mark H. M.; Van Den Herik, H. Jaap (2008). "Parallel
Monte-Carlo Tree Search". Computers and Games. Lecture Notes in Computer Science.
Vol. 5131. pp. 60–71. CiteSeerX 10.1.1.159.4373 (https://citeseerx.ist.psu.edu/viewdoc/sum
mary?doi=10.1.1.159.4373). doi:10.1007/978-3-540-87608-3_6 (https://doi.org/10.1007%2F
978-3-540-87608-3_6). ISBN 978-3-540-87607-6.
81. Bruns, Pete. Monte-Carlo Tree Search in the game of Tantrix: Cosc490 Final Report (https://
www.tantrix.com/Tantrix/TRobot/MCTS%20Final%20Report.pdf) (PDF) (Report).
82. Silver, David; Veness, Joel. "Monte-Carlo Planning in Large POMDPs" (http://www0.cs.ucl.a
c.uk/staff/D.Silver/web/Publications_files/pomcp.pdf) (PDF). 0.cs.ucl.ac.uk. Retrieved
28 October 2017.
83. Lorentz, Richard J. (2011). "Improving Monte–Carlo Tree Search in Havannah". Computers
and Games. Lecture Notes in Computer Science. Vol. 6515. pp. 105–115.
Bibcode:2011LNCS.6515..105L (https://ui.adsabs.harvard.edu/abs/2011LNCS.6515..105L).
doi:10.1007/978-3-642-17928-0_10 (https://doi.org/10.1007%2F978-3-642-17928-0_10).
ISBN 978-3-642-17927-3.
84. Jakl, Tomas. "Arimaa challenge – comparison study of MCTS versus alpha-beta methods"
(http://www.arimaa.com/arimaa/papers/ThomasJakl/bc-thesis.pdf) (PDF). Arimaa.com.
Retrieved 28 October 2017.
85. Szirmay-Kalos 2008.
86. "How the Coast Guard Uses Analytics to Search for Those Lost at Sea" (http://insights.dice.c
om/2014/01/03/how-the-coast-guard-uses-analytics-to-search-for-those-lost-at-sea). Dice
Insights. 3 January 2014.
87. Stone, Lawrence D.; Kratzke, Thomas M.; Frost, John R. "Search Modeling and Optimization
in USCG's Search and Rescue Optimal Planning System (SAROPS)" (http://www.ifremer.fr/
web-com/sar2011/Presentations/SARWS2011_STONE_L.pdf) (PDF). Ifremer.fr. Retrieved
28 October 2017.
88. Carmona, René; Del Moral, Pierre; Hu, Peng; Oudjane, Nadia (2012). "An Introduction to
Particle Methods with Financial Applications". In Carmona, René A.; Moral, Pierre Del; Hu,
Peng; et al. (eds.). Numerical Methods in Finance. Springer Proceedings in Mathematics.
Vol. 12. Springer Berlin Heidelberg. pp. 3–49. CiteSeerX 10.1.1.359.7957 (https://citeseerx.i
st.psu.edu/viewdoc/summary?doi=10.1.1.359.7957). doi:10.1007/978-3-642-25746-9_1 (http
s://doi.org/10.1007%2F978-3-642-25746-9_1). ISBN 978-3-642-25745-2. {{cite
book}}: Missing or empty |title= (help)
89. Carmona, René; Del Moral, Pierre; Hu, Peng; Oudjane, Nadia (2012). Numerical Methods in
Finance (https://basepub.dauphine.fr/handle/123456789/11498). Springer Proceedings in
Mathematics. Vol. 12. doi:10.1007/978-3-642-25746-9 (https://doi.org/10.1007%2F978-3-64
2-25746-9). ISBN 978-3-642-25745-2.
90. Kroese, D. P.; Taimre, T.; Botev, Z. I. (2011). Handbook of Monte Carlo Methods. John Wiley
& Sons.
91. Arenas, Daniel J.; Lett, Lanair A.; Klusaritz, Heather; Teitelman, Anne M. (2017). "A Monte
Carlo simulation approach for estimating the health and economic impact of interventions
provided at a student-run clinic" (https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5746244).
PLOS ONE. 12 (12): e0189718. Bibcode:2017PLoSO..1289718A (https://ui.adsabs.harvard.
edu/abs/2017PLoSO..1289718A). doi:10.1371/journal.pone.0189718 (https://doi.org/10.137
1%2Fjournal.pone.0189718). PMC 5746244 (https://www.ncbi.nlm.nih.gov/pmc/articles/PM
C5746244). PMID 29284026 (https://pubmed.ncbi.nlm.nih.gov/29284026).
92. Elwart, Liz; Emerson, Nina; Enders, Christina; Fumia, Dani; Murphy, Kevin (December
2006). "Increasing Access to Restraining Orders for Low Income Victims of Domestic
Violence: A Cost-Benefit Analysis of the Proposed Domestic Abuse Grant Program" (https://
web.archive.org/web/20181106220526/https://legalaidresearch.org/wp-content/uploads/Res
earch-Increasing-Access-to-REstraining-Order-for-Low-Income-Victims-of-DV-A-Cost-Benefi
t-Analysis-of-the-Proposed-Domestic-Abuse-Grant-Program.pdf) (PDF). State Bar of
Wisconsin. Archived from the original (http://legalaidresearch.org/wp-content/uploads/Resea
rch-Increasing-Access-to-REstraining-Order-for-Low-Income-Victims-of-DV-A-Cost-Benefit-
Analysis-of-the-Proposed-Domestic-Abuse-Grant-Program.pdf) (PDF) on 6 November 2018.
Retrieved 12 December 2016.
93. Dahlan, Hadi Akbar (29 October 2021). "Perbandingan Penerbitan dan Harga Buku
Mengikut Genre di Malaysia dan Jepun Menggunakan Data Akses Terbuka dan Simulasi
Monte Carlo" (http://web.usm.my/km/39(2)2021/KM39022021_8.pdf) (PDF). Kajian
Malaysia. 39 (2): 179–202. doi:10.21315/km2021.39.2.8 (https://doi.org/10.21315%2Fkm202
1.39.2.8). S2CID 240435973 (https://api.semanticscholar.org/CorpusID:240435973).
94. Press et al. 1996
95. MEZEI, M (31 December 1986). "Adaptive umbrella sampling: Self-consistent determination
of the non-Boltzmann bias". Journal of Computational Physics. 68 (1): 237–248.
Bibcode:1987JCoPh..68..237M (https://ui.adsabs.harvard.edu/abs/1987JCoPh..68..237M).
doi:10.1016/0021-9991(87)90054-4 (https://doi.org/10.1016%2F0021-9991%2887%299005
4-4).
96. Bartels, Christian; Karplus, Martin (31 December 1997). "Probability Distributions for
Complex Systems: Adaptive Umbrella Sampling of the Potential Energy". The Journal of
Physical Chemistry B. 102 (5): 865–880. doi:10.1021/jp972280j (https://doi.org/10.1021%2Fj
p972280j).
97. Del Moral, Pierre; Doucet, Arnaud; Jasra, Ajay (2006). "Sequential Monte Carlo samplers".
Journal of the Royal Statistical Society, Series B. 68 (3): 411–436. arXiv:cond-mat/0212648
(https://arxiv.org/abs/cond-mat/0212648). doi:10.1111/j.1467-9868.2006.00553.x (https://doi.
org/10.1111%2Fj.1467-9868.2006.00553.x). S2CID 12074789 (https://api.semanticscholar.o
rg/CorpusID:12074789).
98. Spall, J. C. (2003), Introduction to Stochastic Search and Optimization: Estimation,
Simulation, and Control, Wiley, Hoboken, NJ. http://www.jhuapl.edu/ISSO
99. Mosegaard & Tarantola 1995
100. Tarantola 2005
101. McCracken, D. D., (1955) The Monte Carlo Method, Scientific American, 192(5), pp. 90-97
102. Elishakoff, I., (2003) Notes on Philosophy of the Monte Carlo Method, International Applied
Mechanics, 39(7), pp.753-762
103. Grüne-Yanoff, T., & Weirich, P. (2010). The philosophy and epistemology of simulation: A
review, Simulation & Gaming, 41(1), pp. 20-50

Sources
Anderson, Herbert L. (1986). "Metropolis, Monte Carlo and the MANIAC" (http://library.lanl.g
ov/cgi-bin/getfile?00326886.pdf) (PDF). Los Alamos Science. 14: 96–108.
Benov, Dobriyan M. (2016). "The Manhattan Project, the first electronic computer and the
Monte Carlo method". Monte Carlo Methods and Applications. 22 (1): 73–79.
doi:10.1515/mcma-2016-0102 (https://doi.org/10.1515%2Fmcma-2016-0102).
S2CID 30198383 (https://api.semanticscholar.org/CorpusID:30198383).
Baeurle, Stephan A. (2009). "Multiscale modeling of polymer materials using field-theoretic
methodologies: A survey about recent developments". Journal of Mathematical Chemistry.
46 (2): 363–426. doi:10.1007/s10910-008-9467-3 (https://doi.org/10.1007%2Fs10910-008-9
467-3). S2CID 117867762 (https://api.semanticscholar.org/CorpusID:117867762).
Berg, Bernd A. (2004). Markov Chain Monte Carlo Simulations and Their Statistical Analysis
(With Web-Based Fortran Code). Hackensack, NJ: World Scientific. ISBN 978-981-238-935-
0.
Binder, Kurt (1995). The Monte Carlo Method in Condensed Matter Physics. New York:
Springer. ISBN 978-0-387-54369-7.
Caflisch, R. E. (1998). Monte Carlo and quasi-Monte Carlo methods. Acta Numerica. Vol. 7.
Cambridge University Press. pp. 1–49.
Davenport, J. H. (1992). "Primality testing revisited". Papers from the international
symposium on Symbolic and algebraic computation - ISSAC '92. Proceeding ISSAC '92
Papers from the International Symposium on Symbolic and Algebraic Computation. pp. 123–
129. CiteSeerX 10.1.1.43.9296 (https://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.4
3.9296). doi:10.1145/143242.143290 (https://doi.org/10.1145%2F143242.143290).
ISBN 978-0-89791-489-5. S2CID 17322272 (https://api.semanticscholar.org/CorpusID:1732
2272).
Doucet, Arnaud; Freitas, Nando de; Gordon, Neil (2001). Sequential Monte Carlo methods in
practice. New York: Springer. ISBN 978-0-387-95146-1.
Eckhardt, Roger (1987). "Stan Ulam, John von Neumann, and the Monte Carlo method" (htt
p://library.lanl.gov/cgi-bin/getfile?15-13.pdf) (PDF). Los Alamos Science (15): 131–137.
Fishman, G. S. (1995). Monte Carlo: Concepts, Algorithms, and Applications. New York:
Springer. ISBN 978-0-387-94527-9.
C. Forastero and L. Zamora and D. Guirado and A. Lallena (2010). "A Monte Carlo tool to
simulate breast cancer screening programmes". Phys. Med. Biol. 55 (17): 5213–5229.
Bibcode:2010PMB....55.5213F (https://ui.adsabs.harvard.edu/abs/2010PMB....55.5213F).
doi:10.1088/0031-9155/55/17/021 (https://doi.org/10.1088%2F0031-9155%2F55%2F17%2F
021). PMID 20714045 (https://pubmed.ncbi.nlm.nih.gov/20714045). S2CID 30021759 (http
s://api.semanticscholar.org/CorpusID:30021759).
Golden, Leslie M. (1979). "The Effect of Surface Roughness on the Transmission of
Microwave Radiation Through a Planetary Surface". Icarus. 38 (3): 451–455.
Bibcode:1979Icar...38..451G (https://ui.adsabs.harvard.edu/abs/1979Icar...38..451G).
doi:10.1016/0019-1035(79)90199-4 (https://doi.org/10.1016%2F0019-1035%2879%299019
9-4).
Gould, Harvey; Tobochnik, Jan (1988). An Introduction to Computer Simulation Methods,
Part 2, Applications to Physical Systems. Reading: Addison-Wesley. ISBN 978-0-201-
16504-3.
Grinstead, Charles; Snell, J. Laurie (1997). Introduction to Probability (https://archive.org/det
ails/flooved3489). American Mathematical Society. pp. 10 (https://archive.org/details/flooved
3489/page/n19)–11.
Hammersley, J. M.; Handscomb, D. C. (1975). Monte Carlo Methods. London: Methuen.
ISBN 978-0-416-52340-9.
Hartmann, A.K. (2009). Practical Guide to Computer Simulations (https://web.archive.org/we
b/20090211113048/http://worldscibooks.com/physics/6988.html). World Scientific.
ISBN 978-981-283-415-7. Archived from the original (http://www.worldscibooks.com/physics/
6988.html) on 2009-02-11.
Hubbard, Douglas (2007). How to Measure Anything: Finding the Value of Intangibles in
Business (https://archive.org/details/howtomeasureanyt00hubb_773). John Wiley & Sons.
p. 46 (https://archive.org/details/howtomeasureanyt00hubb_773/page/n64).
ISBN 9780470110126.
Hubbard, Douglas (2009). The Failure of Risk Management: Why It's Broken and How to Fix
It. John Wiley & Sons.
Kahneman, D.; Tversky, A. (1982). Judgement under Uncertainty: Heuristics and Biases.
Cambridge University Press.
Kalos, Malvin H.; Whitlock, Paula A. (2008). Monte Carlo Methods. Wiley-VCH. ISBN 978-3-
527-40760-6.
Kroese, D. P.; Taimre, T.; Botev, Z.I. (2011). Handbook of Monte Carlo Methods (http://www.m
ontecarlohandbook.org). New York: John Wiley & Sons. p. 772. ISBN 978-0-470-17793-8.
MacGillivray, H. T.; Dodd, R. J. (1982). "Monte-Carlo simulations of galaxy systems".
Astrophysics and Space Science. 86 (2): 419–435. doi:10.1007/BF00683346 (https://doi.org/
10.1007%2FBF00683346). S2CID 189849365 (https://api.semanticscholar.org/CorpusID:18
9849365).
MacKeown, P. Kevin (1997). Stochastic Simulation in Physics. New York: Springer.
ISBN 978-981-3083-26-4.
Metropolis, N. (1987). "The beginning of the Monte Carlo method" (http://library.lanl.gov/cgi-b
in/getfile?15-12.pdf) (PDF). Los Alamos Science (1987 Special Issue dedicated to Stanislaw
Ulam): 125–130.
Metropolis, N.; Rosenbluth, Arianna W.; Rosenbluth, Marshall N.; Teller, Augusta H.; Teller,
Edward (1953). "Equation of State Calculations by Fast Computing Machines". Journal of
Chemical Physics. 21 (6): 1087. Bibcode:1953JChPh..21.1087M (https://ui.adsabs.harvard.e
du/abs/1953JChPh..21.1087M). doi:10.1063/1.1699114 (https://doi.org/10.1063%2F1.16991
14). OSTI 4390578 (https://www.osti.gov/biblio/4390578). S2CID 1046577 (https://api.seman
ticscholar.org/CorpusID:1046577).
Metropolis, N.; Ulam, S. (1949). "The Monte Carlo Method". Journal of the American
Statistical Association. 44 (247): 335–341. doi:10.1080/01621459.1949.10483310 (https://do
i.org/10.1080%2F01621459.1949.10483310). JSTOR 2280232 (https://www.jstor.org/stable/
2280232). PMID 18139350 (https://pubmed.ncbi.nlm.nih.gov/18139350).
Milik, M.; Skolnick, J. (Jan 1993). "Insertion of peptide chains into lipid membranes: an off-
lattice Monte Carlo dynamics model" (https://semanticscholar.org/paper/793bd7ab0e505ef5
d12ed2f0798b22675e088407). Proteins. 15 (1): 10–25. doi:10.1002/prot.340150104 (https://
doi.org/10.1002%2Fprot.340150104). PMID 8451235 (https://pubmed.ncbi.nlm.nih.gov/8451
235). S2CID 7450512 (https://api.semanticscholar.org/CorpusID:7450512).
Mosegaard, Klaus; Tarantola, Albert (1995). "Monte Carlo sampling of solutions to inverse
problems" (http://www.math.pitt.edu/~cbsg/Materials/MonteCarlo_latex.pdf) (PDF). J.
Geophys. Res. 100 (B7): 12431–12447. Bibcode:1995JGR...10012431M (https://ui.adsabs.h
arvard.edu/abs/1995JGR...10012431M). doi:10.1029/94JB03097 (https://doi.org/10.1029%2
F94JB03097).
Ojeda, P.; Garcia, M.; Londono, A.; Chen, N.Y. (Feb 2009). "Monte Carlo Simulations of
Proteins in Cages: Influence of Confinement on the Stability of Intermediate States" (https://
www.ncbi.nlm.nih.gov/pmc/articles/PMC2716574). Biophys. J. 96 (3): 1076–1082.
Bibcode:2009BpJ....96.1076O (https://ui.adsabs.harvard.edu/abs/2009BpJ....96.1076O).
doi:10.1529/biophysj.107.125369 (https://doi.org/10.1529%2Fbiophysj.107.125369).
PMC 2716574 (https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2716574). PMID 18849410
(https://pubmed.ncbi.nlm.nih.gov/18849410).
Int Panis, L.; De Nocker, L.; De Vlieger, I.; Torfs, R. (2001). "Trends and uncertainty in air
pollution impacts and external costs of Belgian passenger car traffic". International Journal
of Vehicle Design. 27 (1–4): 183–194. doi:10.1504/IJVD.2001.001963 (https://doi.org/10.150
4%2FIJVD.2001.001963).
Int Panis, L.; Rabl, A.; De Nocker, L.; Torfs, R. (2002). Sturm, P. (ed.). "Diesel or Petrol ? An
environmental comparison hampered by uncertainty". Mitteilungen Institut für
Verbrennungskraftmaschinen und Thermodynamik. Technische Universität Graz Austria.
Heft 81 Vol 1: 48–54.
Press, William H.; Teukolsky, Saul A.; Vetterling, William T.; Flannery, Brian P. (1996) [1986].
Numerical Recipes in Fortran 77: The Art of Scientific Computing. Fortran Numerical
Recipes. Vol. 1 (2nd ed.). Cambridge University Press. ISBN 978-0-521-43064-7.
Ripley, B. D. (1987). Stochastic Simulation. Wiley & Sons.
Robert, C.; Casella, G. (2004). Monte Carlo Statistical Methods (https://archive.org/details/sp
ringer_10.1007-978-1-4757-4145-2) (2nd ed.). New York: Springer. ISBN 978-0-387-21239-
5.
Rubinstein, R. Y.; Kroese, D. P. (2007). Simulation and the Monte Carlo Method (2nd ed.).
New York: John Wiley & Sons. ISBN 978-0-470-17793-8.
Savvides, Savvakis C. (1994). "Risk Analysis in Investment Appraisal" (https://mpra.ub.uni-m
uenchen.de/10035/1/MPRA_paper_10035.pdf) (PDF). Project Appraisal Journal. 9 (1).
doi:10.2139/ssrn.265905 (https://doi.org/10.2139%2Fssrn.265905). S2CID 2809643 (https://
api.semanticscholar.org/CorpusID:2809643).
Sawilowsky, Shlomo S.; Fahoome, Gail C. (2003). Statistics via Monte Carlo Simulation with
Fortran. Rochester Hills, MI: JMASM. ISBN 978-0-9740236-0-1.
Sawilowsky, Shlomo S. (2003). "You think you've got trivials?" (https://digitalcommons.wayn
e.edu/cgi/viewcontent.cgi?article=1744&context=jmasm). Journal of Modern Applied
Statistical Methods. 2 (1): 218–225. doi:10.22237/jmasm/1051748460 (https://doi.org/10.222
37%2Fjmasm%2F1051748460).
Silver, David; Veness, Joel (2010). "Monte-Carlo Planning in Large POMDPs" (http://books.n
ips.cc/papers/files/nips23/NIPS2010_0740.pdf) (PDF). In Lafferty, J.; Williams, C. K. I.;
Shawe-Taylor, J.; Zemel, R. S.; Culotta, A. (eds.). Advances in Neural Information
Processing Systems 23. Neural Information Processing Systems 2010. Neural Information
Processing Systems Foundation.
Szirmay-Kalos, László (2008). Monte Carlo Methods in Global Illumination - Photo-realistic
Rendering with Randomization. VDM Verlag Dr. Mueller e.K. ISBN 978-3-8364-7919-6.
Tarantola, Albert (2005). Inverse Problem Theory (http://www.ipgp.jussieu.fr/~tarantola/Files/
Professional/SIAM/index.html). Philadelphia: Society for Industrial and Applied Mathematics.
ISBN 978-0-89871-572-9.
Vose, David (2008). Risk Analysis, A Quantitative Guide (3rd ed.). John Wiley & Sons.
ISBN 9780470512845.
Mazhdrakov, Metodi; Benov, Dobriyan; Valkanov, Nikolai (2018). The Monte Carlo Method.
Engineering Applications (https://books.google.com/books?id=t0BqDwAAQBAJ&q=the+mo
nte+carlo+method+engineering+applications+mazhdrakov). ACMO Academic Press.
ISBN 978-619-90684-3-4.

External links
Media related to Monte Carlo method at Wikimedia Commons

Retrieved from "https://en.wikipedia.org/w/index.php?title=Monte_Carlo_method&oldid=1166551287"

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy