Math 183 Spring 2012 Notes
Math 183 Spring 2012 Notes
Math 183 Spring 2012 Notes
Adolfo J. Rumbos
c Draft date March 23, 2012
2
Contents
1 Preface 5
2 Modeling Process 7
2.1 Constructing Models . . . . . . . . . . . . . . . . . . . . . . . . . 7
2.1.1 Conservation Principles . . . . . . . . . . . . . . . . . . . 8
2.1.2 Constitutive Equations . . . . . . . . . . . . . . . . . . . . 8
2.2 Example: Bacterial Growth in a Chemostat . . . . . . . . . . . . 8
2.3 Analysis of Models . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.3.1 Nondimensionalization . . . . . . . . . . . . . . . . . . . . 11
4 Stochastic Models 35
4.1 Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . 36
4.1.1 A Brief Excursion into Probability . . . . . . . . . . . . . 37
4.1.2 Discrete Random Variables . . . . . . . . . . . . . . . . . 40
4.1.3 The Binomial Distribution . . . . . . . . . . . . . . . . . . 42
4.1.4 Expected Value . . . . . . . . . . . . . . . . . . . . . . . . 48
4.1.5 The Poisson Distribution . . . . . . . . . . . . . . . . . . 49
4.1.6 Estimating Mutation Rates in Bacterial Populations . . . 52
4.1.7 Another Brief Excursion into Probability . . . . . . . . . 53
4.1.8 Continuous Random Variables . . . . . . . . . . . . . . . 56
4.2 Random Processes . . . . . . . . . . . . . . . . . . . . . . . . . . 61
3
4 CONTENTS
Chapter 1
Preface
The main goal of this course is to provide opportunities for students to con-
struct and analyze mathematical models that arise in the physical, biological
and social sciences. Mathematical models are usually created in order to obtain
understanding of problems and situations arising in the real world; other times,
the main goal is to make predictions or to control certain processes; finally, some
models are created in order to aid in decision making.
Construction of a mathematical model consists of translating a real world
problem into a mathematical problem involving parameters, variables, functions,
equations and/or inequalities. Analysis of the model involves the solution (if
possible) of the mathematical problem through logical, algebraic, analytical or
computational means, and assessing what the solutions imply about the real
situation under study. If an analytical or computational solution is not possible,
computer simulations can sometimes be used in order to study various scenarios
implied or predicted by the model.
Analysis techniques can be drawn from many areas of mathematics. In
this course, it is assumed that students have a good working knowledge of
Calculus, Linear Algebra and Ordinary Differential Equations. These areas are
adequate for the analysis of some models. However, many modeling situations
require the use of some probability theory and linear programming. These
mathematical topics will be covered in the course. In calculus and differential
equations courses, students have been exposed to some continuous models. In
this course, we will also introduce students to discrete models as well.
5
6 CHAPTER 1. PREFACE
Chapter 2
We will get a chance to go through at least the first two stages listed above in
a variety of case studies or examples.
Of course, the modeling process always begins with a question that we want
to answer, or a problem we have to solve. Often, asking the right questions and
posing the right problems can be the hardest part in the modeling process. This
part of the process involves getting acquainted with the intricacies of the science
involved in the particular question at hand. It is unrealistic to expect that a
mathematical modeling course will teach students to do this in a systematic way.
The best we can do is to present many case studies and examples of real life
modeling situations that mathematicians have analyzed in various situations.
One of the goals of the course is to have students grapple with this issue when
working with a specific problem in a term project that will involve a large portion
of the course.
7
8 CHAPTER 2. MODELING PROCESS
dQ
= Rate of Q in − Rate of Q out. (2.1)
dt
In this case, the conservation principle might lead to a differential equation, or
a system of differential equations, and so the theory of differential equations can
be used to help in the analysis of the model.
co
F -
N (t)
Q(t)
F
-
dN
= Rate of N in − Rate of N out. (2.5)
dt
We are assuming here that N is a differentiable function of time. This assump-
tion is justified if
(i) we are dealing with populations of very large size so that the addition (or
removal) of a few individuals is not very significant; for example, in the
case of a bacterial colony, N is of the order of 106 cells per milliliter;
(ii) ”there are no distinct population changes that occur at timed intervals,”
see [EK88, pg. 117].
Using the constitutive assumption stated previously, we have that
where n is the bacterial density defined in (2.4). We can therefore re–write (2.5)
as
dN F
= K(c)N − N. (2.8)
dt V
Next, apply the conservation principle (2.1) to the amount of nutrient, Q(t),
in the chamber, where
Rate of Q in = F co , (2.9)
and
Rate of Q out = F c + αK(c)N, (2.10)
where we have introduced another parameter α, which measures the fraction of
nutrient that is being consumed as a result of bacterial growth. The reciprocal
of the parameter α,
1
Y = , (2.11)
α
measures the number of cells produced due to consumption of one unit of nu-
trient, and is usually referred to as the yield.
2.3. ANALYSIS OF MODELS 11
Combining (2.10), (2.9) and (2.1) we see that the conservation principle for
Q takes the form
dQ
= F co − F c − αK(c)N. (2.12)
dt
Using the definition of c in (2.3) we can re–write (2.12) as
dQ F
= F co − Q − αK(c)N. (2.13)
dt V
The differential equations in (2.8) and (2.13) yield the system of differential
equations
dN F
dt = K(c)N − V N ;
(2.14)
dQ F
= F co − Q − αK(c)N.
dt V
Thus, application of conservation principles and a few constitutive assumptions
has yielded a system of ordinary differential equations (2.14) for the variables N
and Q in the chemostat system. We have therefore constructed a preliminary
mathematical model for bacterial growth in a chemostat.
Dividing the equations in (2.14) by the fixed volume, V , of the culture in
the chamber, we obtain the following system of ordinary differential equations
for the bacterial population density, n(t), and the nutrient concentration, c(t).
dn F
dt = K(c)n − V n;
(2.15)
dc F F
= co − c − αK(c)n.
dt V V
Thus, we have arrived at a mathematical model that describes the evolution
in time of the population density and nutrient concentration in a chemostat
system. We will analyze the system in (2.15) in subsequent sections.
2.3.1 Nondimensionalization
In this section we illustrate yet another way to simplify the problem which
consists of introducing dimensionless variables (variables without units). This
12 CHAPTER 2. MODELING PROCESS
b
2
a c
get
c
c=
b . (2.18)
a
It is not clear at the moment what the scaling factor for n and t should be, so
we shall denote them by µ and λ, respectively. We then have that,
n
n
b= , (2.19)
µ
and
t
τ= , (2.20)
λ
where µ has units of bacterial density (cells/volume), and λ has units of time.
Next, we find expressions for the derivatives
db
n db
c
and . (2.21)
dτ dτ
To find the expressions in (2.21) we need to apply the Chain Rule; for instance,
db
n db
n dt
= · (2.22)
dτ dt dτ
To compute the right–hand side of (2.22), we use (2.19) and (2.20) to obtain
from (2.22) that
db
n λ dn
= . (2.23)
dτ µ dt
dn
Next, substitute the expression for in the first equation in (2.17) into the
dt
right–hand side of (2.23) to obtain
db
n λ bnb
c F
= − n , (2.24)
dτ µ 1+bc V
where we have also used the expression for b
c in (2.18). Distributing on the
right–hand side of (2.24) we obtain
db
n n c λF
−
bb
= λb n
b, (2.25)
dτ 1+b c V
where we have used (2.19).
We will now choose λ so that
λF
= 1, (2.26)
V
from which we get that
V
λ= (2.27)
F
is our scaling factor for t; observe that the parameter λ in (2.27) has units of
time.
14 CHAPTER 2. MODELING PROCESS
db
n n c
−n
bb
= α1 b. (2.30)
dτ 1+b c
Similar calculations (see Problem 1 in Assignment 2) show that
db
c n c
= α2 − −b
bb
c, (2.31)
dτ 1+b c
where we have set
co
α2 = (2.32)
a
and
αbλµ
= 1,
a
so that
a
µ= . (2.33)
αbλ
Note that the parameter α2 in (2.32) is dimensionless and that that the units
of µ defined in (2.33) are cells/volume.
Putting together the equations in (2.30) and (2.33) we obtain the system
dbn n c
−n
bb
dτ = α1 1 + b b;
c
(2.34)
dbc n c
= α2 − −b
b b
c,
dτ 1+b c
in the nondimensional variables n
b, b
c and τ defined in (2.19), (2.18) and (2.20),
respectively. Observe that the system in (2.34) contains two dimensionless pa-
rameters, α1 and α2 , as opposed to the six parameters in the original system
in (2.17). This reduction in the number of parameters greatly simplifies the
problem in two aspects:
For instance, the equilibrium points of the system in (2.34) are expressed in
terms of the parameters α1 and α2 as follows
1 1
(0, α2 ) and α1 α2 − , . (2.35)
α1 − 1 α1 − 1
F < bV
and
aF
co > ,
bV − F
respectively.
The equilibrium solution (0, α2 ) in (2.35) is referred to as the “washout”
solution, since all the bacteria washed out because of the flow; while the second
solution in (2.35) is the “survival” solution.
Stability analysis of the dimensionless system in (2.34) will reveal further
conditions that determine whether the chemostat system will yield a sustainable
crop of bacteria. Some of this analysis is carried out in Assignment 2.
16 CHAPTER 2. MODELING PROCESS
Chapter 3
Continuous Deterministic
Models
17
18 CHAPTER 3. CONTINUOUS DETERMINISTIC MODELS
v1 r - v2 r -
x=a x=b
dN
= Number of cars entering at a − Number of cars leaving at b. (3.3)
dt
We can re–write the conservation principle in (3.3) more succinctly by postulat-
ing a traffic flux function, q(x, t), which measures the number of cars crossing
location x per unit of time at time t. Using the traffic flux function we can then
re–write (3.3) as
dN
= q(a, t) − q(b, t),
dt
or
dN
= −[q(b, t) − q(a, t)]. (3.4)
dt
Assuming that the flux function is differentiable and that its partial derivatives
are continuous, we can invoke the Fundamental Theorem of Calculus to re–write
(3.4) as
Z b
dN ∂
=− [q(x, t)] dx. (3.5)
dt a ∂x
3.1. EXAMPLE: MODELING TRAFFIC FLOW 19
Next, assume that the traffic density, ρ, has continuous partial derivatives
to obtain from (3.2) that
Z b
dN ∂
= [ρ(x, t)] dx. (3.6)
dt a ∂t
Combining (3.6) and (3.5) we then see that the conservation principle in (3.4)
now takes the form
Z b Z b
∂ ∂
[ρ(x, t)] dx = − [q(x, t)] dx. (3.7)
a ∂t a ∂x
Rewrite the equation in (3.7) as
Z b Z b
∂ ∂
[ρ(x, t)] dx + [q(x, t)] dx = 0,
a ∂t a ∂x
or Z b
∂ ∂
[ρ(x, t)] + [q(x, t)] dx = 0, (3.8)
a ∂t ∂x
and observe that the points a and b are arbitrary. Since, we are assuming that
the partial derivatives of ρ and q are continuous, we can show (see Assignment
4) that, given that (3.8) holds true for all intervals [a, b], then we must have
that
∂ ∂
[ρ(x, t)] + [q(x, t)] = 0. (3.9)
∂t ∂x
The equation in (3.9) is an example of a partial differential equation or
PDE. It is usually written in a more compact form
∂ρ ∂q
+ = 0, (3.10)
∂t ∂x
or
ρt + qx = 0, (3.11)
where the subscripts in (3.11) indicate partial derivatives with respect to the
subscripted variables.
Ideally, we would like to find a solution, ρ, to (3.10) subject to some initial
condition
ρ(x, 0) = ρo (x), (3.12)
for some initial traffic density profile, ρo , along the road.
Before we proceed any further, we need to model the traffic flux, q(x, t).
Imagine a vehicle at x is moving with a velocity v. Then, for a short time
interval, [t, t + ∆t] the car moves a distance approximately give by
∆x ≈ v∆t.
The number of cars in the section [x, x + ∆x] is then, approximately, given by
q = vρ. (3.14)
Later in this course we shall shall see how to derive expressions like (3.17)
relating traffic velocity to traffic density from theoretical considerations; for
now, we simply note that it models the intuitive notion that we stated above
regarding the traffic velocity being low for high traffic density.
The partial differential equation model for traffic flow presented in this sec-
tion, based on the conservation equation in (3.15) and a constitutive relation
for the traffic velocity, v, and the traffic density ρ (of which (3.17) is just an
example), was first introduced by Lighthill and Whitman in 1955 (see [LW55]);
it was also treated by Richards in 1956, [Ric56].
1 ∂ρ vmax ∂
+ [(1 − u) ρ] = 0. (3.23)
τ ∂b
t L ∂bx
Next, divide the equation in (3.23) and use the first equation in (3.20) to get
∂u vmax τ ∂
+ [(1 − u) u] = 0, (3.24)
∂b
t L ∂b x
where we have also multiplied by τ . Setting
vmax τ
= 1,
L
or, equivalently, choosing the time scale τ to be L/vmax , we see that we can
re–write (3.24) as
∂u ∂
+ [(1 − u) u] = 0. (3.25)
∂tb ∂b
x
The equation in (3.25) is now in dimensionless form. If the original time and
space variables, t and x, are assumed to be in units of τ and L, respectively, we
can then rewrite the equation in (3.25) as
∂u ∂
+ [(1 − u) u] = 0, (3.26)
∂t ∂x
where u, x and t represent real (dimensionless) variables. The equation in (3.26)
is the one we’ll be analyzing for the remainder of this section.
Set
g(u) = u(1 − u); (3.27)
22 CHAPTER 3. CONTINUOUS DETERMINISTIC MODELS
then, the partial differential equation in (3.26) can be written in the form
∂u ∂
+ [g(u)] = 0, (3.28)
∂t ∂x
or
∂u ∂u
+ g 0 (u) = 0. (3.29)
∂t ∂x
We will describe here a general procedure for analyzing the equation in (3.28)
for the case in which g is a differentiable function. We begin by presenting the
simplest example of a linear function
g(u) = cu,
for some constant c. The equation in (3.29) then becomes the linear first order
partial differential equation
∂u ∂u
+c = 0. (3.30)
∂t ∂x
We will show how to find a solution to the differential equation in (3.30) subject
to the initial condition
u(x, 0) = f (x), (3.31)
where f is some differentiable function of a single variable.
The problem of determining a solution of the differential equation in (3.30)
subject to the condition in (3.31) is an example of initial value problem and is
usually written in a the more compact form
∂u ∂u
+c = 0, x ∈ R, t > 0;
∂t ∂x (3.32)
x ∈ R.
u(x, 0) = f (x),
Differentiate with respect to t the function defined in (3.34), using the Chain
Rule, to obtain
d ∂u dx ∂u dt
[u(x(t), t)] = + ,
dt ∂x dt ∂t dt
3.2. ANALYSIS OF THE TRAFFIC FLOW EQUATION 23
du
= 0. (3.37)
dt
The ordinary differential equation in (3.36) defines a family of curves in the
xt–plane give by
x = ct + k, (3.38)
where k is a real parameter. The curves in (3.38) are straight lines of slope 1/c
in the xt–plane; some of the curves for the case c > 0 are pictured in Figure
3.2.3.
t
x
The curves in (3.38) are called the characteristic curves of the partial dif-
ferential equation in (3.32) and are defined by the ordinary differential equation
in (3.36).
Since the equation in (3.37) was derived by differentiating u along a char-
acteristic curve, it implies that u is constant along characteristics. We can
therefore conclude from (3.37) that
The equation in (3.39) allows us to obtain a formula for u(x, t), where (x, t) lies
along the characteristic indexed by k in (3.38) as follows
where ϕ(k) denotes the constant value of u along the characteristic in (3.38)
indexed by k.
Solving for k in (3.38) and substituting into (3.40) yields a general formula
for computing a solution to the partial differential equation in (3.32):
so that
u(x, t) = f (x − ct), for x ∈ R, t ∈ R, (3.42)
solves the initial value problem (3.32). The expression in (3.42) says that the
solution to (3.32) is a traveling wave, which moves to the right with speed c
if c > 0, or to the left if c < 0. In other words, the initial profile, f (x), moves
propagates without distortion with velocity c. The solution in (3.42) is also
known as an advection wave and the partial differential equation in (3.32) is
known as the advection equation.
Example 3.2.2 (Inviscid Burgers’ Equation). Solve the equation in (3.44) sub-
ject to the initial condition
u(x, 0) = f (x),
where f is some given continuous function; in other words, solve the initial value
problem
∂u ∂u
+u = 0, x ∈ R, t > 0;
∂t ∂x (3.45)
x ∈ R.
u(x, 0) = f (x),
We proceed as in Example 3.2.1 by first obtaining the equation for the charac-
teristic curves
dx
= u. (3.46)
dt
3.2. ANALYSIS OF THE TRAFFIC FLOW EQUATION 25
In this case we cannot solve directly for the characteristic curves. However, as
in Example 3.2.1, the a solution, u, to the partial differential equation in (3.45)
must solve the ordinary differential equation
du
= 0; (3.47)
dt
Thus, u must be constant along the characteristic curves given by (3.46). Thus,
we can solve (3.47) to yield
u(x, t) = ϕ(k), (3.48)
where ϕ(k) denotes the constant value of u along the characteristic curve given
by (3.46) indexed by k. We can then re–write (3.46) as
dx
= ϕ(k),
dt
which can be solved to yield the characteristic curves
x = ϕ(k)t + k. (3.49)
Example 3.2.3 (Inviscid Burgers’ Equation, Continued). Solve the initial value
problem in (3.45) where the initial condition is given by
Example 3.2.4 (Inviscid Burgers’ Equation, Continued). Solve the initial value
problem in (3.45) where the initial condition is given by
Figure 3.2.4 shows a sketch of the initial condition, f . The characteristic curves
f
for the differential equation in (3.45) are solutions to the ordinary differential
equation
dx
= u. (3.59)
dt
Along the characteristic curves, u solves the ordinary differential equation
du
= 0, (3.60)
dt
3.2. ANALYSIS OF THE TRAFFIC FLOW EQUATION 27
x = ϕ(k)t + k. (3.62)
Next, solve for k in (3.62) and substitute in (3.61) to obtain that u is given
implicitly by
u(x, t) = ϕ(x − u(x, t)t). (3.63)
so that
u(x, t) = f (x − u(x, t)t). (3.65)
It follows from (3.62) and (3.64) that the characteristic curves are given by the
equations
x = f (k)t + k; (3.66)
since f (k) = 0 for k 6 0; for 0 < k < 1, the characteristic curves are the straight
lines
x = k(t + 1);
and, for k > 1, the characteristic lines are the straight lines of slope 1
x = t + k.
A sketch of The characteristic curves is shown in Figure 3.2.5. Notice that the
characteristic curves for 0 6 k 6 1 fan out from the t–axis to the line x = t + 1.
Since the solution, u, to the initial value problem in (3.45) is constant along
characteristic curves (see the equations in (3.60) and (3.61)), the sketch in Figure
3.2.5 shows that u can be computed by traveling back along the characteristic
curves to the initial time, t = 0, and reading off value of f (k) for the particular
value point k on the x–axis. Thus, in theory, the initial value problem in (3.45)
with initial condition given in (3.58) can be solved, and the solution is unique.
28 CHAPTER 3. CONTINUOUS DETERMINISTIC MODELS
t
x
Figure 3.2.5: Sketch of the characteristic curves for (3.45) with f given in (3.58)
The fanning our of the characteristic curves pictured in Figure 3.2.5 has the
effect of stretching the initial profile for u in the x direction. This is shown in
Figure 3.2.6, where sketch of u(x, t) as given in (3.67) is shown for t = 0 and
t = 1; the initial profile is shown in dashed lines.
u
x
The nature of the solutions to the initial value problem in (3.45) changes
dramatically when the following initial profile is used.
1,
if x < 0;
f (x) = 1 − x, if 0 6 x < 1; (3.68)
0, if x > 1.
@
@
@
@
@
@
x
Example 3.2.6 (Inviscid Burgers’ Equation, Continued). Solve the initial value
problem in (3.45) where f is as given in (3.68).
Proceeding as in Example 3.2.6, we sketch the characteristic curves in Figure
3.2.8. In Figure 3.2.8 we see that the characteristic curves,
t
x
Figure 3.2.8: Sketch of the characteristic curves for (3.45) with f given in (3.68)
x = f (k)t + k, (3.69)
for 0 < k < 1, instead of fanning out, bunch in and all meet at the single point
with coordinates (1, 1) in the xt–plane. To see why this is the case, take two of
the characteristic curves in (3.69) with equations
x = f (k1 )t + k1 , (3.70)
and
x = f (k2 )t + k2 , (3.71)
with 0 < k1 < k2 6 1. To find the intersection of the lines in (3.70) and (3.71),
set the equations equal to each other and use the definition of f in (3.68), so
that f (k) = 1 − k, for 0 < k 6 1, to get that
(1 − k1 )t + k1 = (1 − k2 )t + k2 ,
30 CHAPTER 3. CONTINUOUS DETERMINISTIC MODELS
from which we get that t = 1. Thus, u(x, t) ceases to exist in the usual sense at
t = 1.
As in Example 3.2.5, we can obtain a formula for computing u(x, t), at least
for t < 1:
1,
if x < t < 1;
1 − x
u(x, t) = , if t < x 6 1; (3.72)
1−t
0, if x > 1, t < 1.
Figure 3.2.9 shows a picture of the graph of u(x, t), for t = 1/2, as a function
x. As t approaches 1 from the left, we see from (3.72) that the profile of u(x, t)
u
A
A
A
A
A
A
x
discontinuity.
d b
Z
u(x, t) dx = Flux at a − Flux at b, (3.73)
dt a
for all a, b ∈ R with a < b. In the case of Burgers’ equation, the flux is given by
1
F (x, t) = [u(x, t)]2 (3.74)
2
Combining (3.73) and (3.74), the equation in (3.73) can be written as
Z b
d 1 1
u(x, t) dx = [u(a, t)]2 − [u(b, t)]2 , (3.75)
dt a 2 2
for all a, b ∈ R with a < b.
The conservation principle in (3.75) can be used to describe what happens
to the solution after a shock forms; for instance, after t = 1 in Example 3.2.6.
We saw in that example that a discontinuity develops. The discontinuity will
continue to travel along some curve in the xt–plane parametrized by a path of
the form
t 7→ (σ(t), t). (3.76)
We would like to describe the path in (3.76). In order to do this, we assume that
the path in (3.76) is differentiable with continuous derivative (in other words,
the path in (3.76) is a C 1 path). We also assume that u has a jump discontinuity
along the path in (3.76), so that the one sided limits
For instance, in Example 3.2.6 we saw that the solution to the inviscid Burgers’
equation with the initial condition in (3.68) has a jump discontinuity at t = 1
with
u− (1) = 1 and u+ (1) = 0. (3.78)
The conservation expression in (3.73) states that the quantity
Z b
Q(a, b, t) = u(x, t) dx, (3.79)
a
since u(x, t + ∆t) is to the left of the shock for ∆t > 0 (see Figure 3.2.11).
Similarly, we get from (3.79) and (3.77) that
t + ∆t
u−
t
u+
x
σ(t) σ(t + ∆t)
Figure 3.2.11: Path of a shock solution for (3.45) with f given in (3.68)
t 7→ (σ(t), t)
dσ 1
= ,
dt 2
so that
t
σ(t) = + c. (3.88)
2
Since σ(1) = 1, we get from (3.88) that
1
c=
2
in (3.88). We therefore get the following formula for the weak solution to initial
value problem discussed in Example 3.2.6:
t+1
1, if x <
, t > 1;
2
u(x, t) = (3.89)
0, if x > t + 1
, t > 1.
2
Figure 3.2.12 shows a picture of some of the characteristic curves for (3.45)
with f given in (3.68), which also incorporates the shock wave solution that we
derived in (3.89).
34 CHAPTER 3. CONTINUOUS DETERMINISTIC MODELS
x
Figure 3.2.12: Sketch of the characteristic curves for (3.45) with f given in
(3.68) with shock wave solution
Chapter 4
Stochastic Models
The models discussed in the previous chapter have been deterministic—the vari-
ables used in the models are completely determined by the values of a set of
parameters and the values of the variables at some initial point or curve. For
instance, when modeling bacterial growth, the number of bacteria in a culture
at time, t, might be modeled by a continuous variable, N (t), which satisfies the
initial value problem
dN N
= rN 1 − ;
dt K
(4.1)
N (0) = No ,
known as the logistic model. The model in (4.1) was derived in Assignment
#1 as a special case of the bacterial growth in a chemostat model presented
in Section 2.2 of these notes. The parameters in the equation in (4.1) are
the intrinsic growth rate, r, and the carrying capacity of the growth medium,
K. Given values for the parameters r and K, and the size of the bacterial
population, No , at time t = 0, the values of population size, N (t), for later
values of t, are completely determined by the formula
No K
N (t) = , for t > 0. (4.2)
No + (K − No )e−rt
35
36 CHAPTER 4. STOCHASTIC MODELS
Example 4.1.1. Suppose that two bacteria, a and b, can randomly develop a
mutation in a unit of time. Assume that each bacterium can mutate at most
once in the unit of time of the experiment. Let M denote the number of bacteria
out of the two that develop mutations after one unit of time. Then M can take
on the values 0, 1, or 2. We cannot predict precisely what value M will take on.
Any time we run the experiment of placing the two bacteria under observation
and counting the number of mutations we may get any of the possible values.
M is thus an example of a random variable. The best we can hope for is an
estimate of the probabilities that M can take on any of the possible values; in
symbols, we want to estimate
We assume that the event that A occurs will not affect the probability of event
B. We say that A and B are stochastically independent.
A ∩ B = [M = 2].
die are equally likely; thus, the probability of any given face is 1/6.
38 CHAPTER 4. STOCHASTIC MODELS
Thus, in Example 4.1.1, Ac is the event that bacterium a does not develop
a mutation in one unit of time. Observe that A and Ac are mutually exclusive;
that is, if A occurs then Ac cannot occur.
Definition 4.1.4 (Mutually Exclusive Events). Events A and B are said to be
mutually exclusive if and only if A ∩ B = ∅.
Definition 4.1.5. Given events A and B, the symbol A ∪ B denotes the event
that either A or B occurs.
Definition 4.1.6 (Probability of Mutually Exclusive Events). If A and B are
mutually exclusive, then
Definition 4.1.8 (Set Difference). Given any events A and B, we define the
set
A\B = {x ∈ A | x 6∈ B}.
Note that A\B and B are mutually exclusive, by Definition 4.1.8. Further-
more,
A = B ∪ (A\B). (4.9)
We therefore have the following proposition.
where
x ∈ A\(A ∩ B) iff x ∈ A and x 6∈ A ∩ B
iff x ∈ A and x 6∈ B
iff x ∈ A ∩ Bc,
so that
A\(A ∩ B) = A ∩ B c . (4.13)
Substituting (4.13) into (4.12) then yields
Observing that
A ∪ B = (A ∩ B c ) ∪ (B ∩ Ac ) ∪ (A ∩ B), (4.16)
Pr(Ac ) = 1 − p, (4.18)
where we have used (4.8) and (4.4). Likewise, the probability that bacterium b
will not mutate is
Pr(B c ) = 1 − p. (4.19)
Since Ac and B c are independent, it follows from (4.5) and (4.19) that
where the events A ∩ B c and Ac ∩ B are mutually exclusive. It then follows from
(4.7) and (4.21) that
Next, use the independence of the events A and B c (see Problem 1 in Assignment
#7) to compute
P (A ∩ B c ) = Pr(A) · Pr(B c )
= p(1 − p),
where we have used Definition 4.1.2, Proposition 4.1.7 and (4.4). Similarly,
[M = 0] = Ac ∩ B c ;
Observe that
2
X
pM (k) = (1 − p)2 + 2(1 − p)p + p2
k=0
= [(1 − p) + p]2
= 1,
so that the function defined in (4.29) satisfies the condition (4.28) in the defini-
tion of a the distribution of a random variable (Definition 4.1.12).
Definition 4.1.14 (Bernoulli Trials). A random experiment with two mutually
exclusive outcomes, one called a “success” and the other a “failure,” is called a
Bernoulli trial. We associate a random variable, X, with a Bernoulli as follows:
X = 1 if the outcome is a success, and X = 0 if the outcome is a failure. If the
probability of a success is p, then then distribution of X is
1 − p if k = 0;
pX (k) = p if k = 1; (4.30)
0 elsewhere.
for all values of a and b; that is, X1 , X2 and X3 are pairwise stochastically
independent, and
Proof: Compute
Pr(Y2 = w, X3 = z) = Pr(X1 + X2 = w, X3 = z)
X
= Pr(X1 = x, X2 = w − x, X3 = z),
x
where the summation is taken over all possible values of X1 . Thus, using (4.34)
in Definition 4.1.18,
X
Pr(Y2 = w, X3 = z) = Pr(X1 = x) · Pr(X2 = w − x) · Pr(X3 = z)
x
!
X
= Pr(X1 = x) · Pr(X2 = w − x) · Pr(X3 = z)
x
= Pr(X1 + X2 = w) · Pr(X3 = z)
Y3 = Y2 + X3 ,
Pr(Y3 = 0) = Pr(Y2 = 0, X3 = 0)
= Pr(Y2 = 0) · Pr(X3 = 0), by independence (Lemma 4.1.19),
= (1 − p)2 · (1 − p)
= (1 − p)3 .
44 CHAPTER 4. STOCHASTIC MODELS
Next, since the event (Y3 = 1) consists of the disjoint union of the events
(Y2 = 1, X3 = 0) and (Y2 = 0, X3 = 1),
where we have used Lemma 4.1.19 and the definition of the probability distri-
bution of Y2 in (4.32). Similarly,
and
Pr(Y3 = 3) = Pr(Y2 = 2, X3 = 1)
= Pr(Y2 = 0) · Pr(X3 = 0)
= p2 · p
= p3 .
We then have that the probability distribution of Y3 is given by
(1 − p)3 if k = 0,
2
3p(1 − p) if k = 1,
2
pY3 (k) = 3p (1 − p) if k = 2, (4.35)
p3 if k = 3,
0 elsewhere.
If we go through the calculations in Examples 4.1.11 and 4.1.20 for the case of
four mutually independent2 Bernoulli trials with parameter p, where 0 < p < 1,
X1 , X2 , X3 and X4 , we obtain that for Y4 = X1 + X2 + X3 + X4 ,
(1 − p)4 if k = 0,
4p(1 − p)3 if k = 1,
6p2 (1 − p)2
if k = 2,
pY4 (k) = (4.36)
4p3 (1 − p) if k = 3,
p 4 if y = 4,
0 elsewhere.
2 Here, not only do we require that the random variable be pairwise independent, but also
that for any group of k ≥ 2 events (Xj = xj ), the probability of their intersection is the
product of their probabilities.
4.1. RANDOM VARIABLES 45
Observe that the terms in the expressions for pY2 (y), pY3 (y) and pY4 (y) in (4.32),
(4.35) and (4.36), respectively, are the terms in the expansion of [(1 − p) + p]n
for n = 2, 3 and 4, respectively. By the Binomial Expansion Theorem,
n
n
X n k n−k
(a + b) = a b , for a, b ∈ R, n ∈ N, (4.37)
k
k=0
where
n n!
= , k = 0, 1, 2 . . . , n, (4.38)
k k!(n − k)!
are the called the binomial coefficients, we obtain that
n
X n k
[(1 − p) + p]n = p (1 − p)n−k .
k
k=0
Yn = X1 + X2 + · · · + Xn .
and
pY1 (1) = Pr(X1 = 1) = p.
Thus, (
1−p if k = 0,
pY1 (k) =
p if k = 1.
46 CHAPTER 4. STOCHASTIC MODELS
1 1
Observe that = = 1 and therefore the result in (4.39) holds true for
0 1
n = 1.
Next, assume the theorem is true for n; that is, suppose that
n k
pYn (k) = p (1 − p)n−k for k = 0, 1, 2, . . . , n, (4.40)
k
and pYn (k) = 0 elsewhere. We show then show that the result also holds true
for n + 1. In other words, we show that if X1 , X2 , . . . , Xn , Xn+1 are mutually
independent Bernoulli trials with parameter p, with 0 < p < 1, and
Yn+1 = Yn + Xn+1 ,
= Pr(Yn = k) · Pr(Xn+1 = 0)
+ Pr(Yn = k − 1) · Pr(Xn−1 = 1)
n k
= p (1 − p)n−k (1 − p)
k
n
+ pk−1 (1 − p)n−k+1 p,
k−1
where we have used the inductive hypothesis (4.40). Thus,
n n
Pr(Yn+1 = k) = + pk (1 − p)n+1−k . (4.43)
k k−1
The expression in (4.42) will following from (4.43) the fact that
n n n+1
+ = , (4.44)
k k−1 k
which can be established by the following counting argument:
4.1. RANDOM VARIABLES 47
= Pr(Yn = n) · Pr(Xn+1 = 1)
= pn p
= pn+1
n+1 k
= p (1 − p)n+1−k ,
k
since k = n + 1, and so (4.42) is established for k = n + 1.
The proof is now complete.
Definition 4.1.22 (Binomial Distribution). A discrete random variable, Y ,
which counts the number of successes in n independent Bernoulli(p) trials, and
having the distribution
n!
k n−k
k!(n − k)! p (1 − p) for k = 0, 1, 2, . . . , n,
pY (k) = (4.47)
0 elsewhere.
is called a binomial random variable with parameters n and p, where p is the
provability of each success. We write
Y ∼ Binomial(n, p).
48 CHAPTER 4. STOCHASTIC MODELS
is called the expected value of X and is denoted by E(X). We then have that
n
X
E(X) = xk pX (xk ) (4.49)
k=1
X1 , X2 , · · · , Xn
where we have used (4.50). Thus, the expected value of a binomial random
variable, Y , with parameters n and p is
Thus, applying the formula for the binomial theorem in (4.37) for n − 1 in place
of n and a = p, b = 1 − p, we obtain from the result in (4.53) that
E(Y ) = np · (p + 1 − p)n−1 ,
n!
k n−k
k!(n − k)! p (1 − p) for k = 0, 1, 2, . . . , n,
pYn (k) = (4.54)
0 elsewhere,
where λ is a constant.
Note that from
np = λ,
we get that
λ
p= . (4.57)
n
50 CHAPTER 4. STOCHASTIC MODELS
p→0 as n → ∞,
λ
where we have used (4.57) to replace p by . Next, rewrite (4.58) as
n
−k n
λk n(n − 1)(n − 2) · · · (n − k + 1)
λ λ
pYn (k) = 1 − 1 − ,
k! nk n n
λk −λ
lim pYn (k) = e . (4.63)
n→∞ k!
4.1. RANDOM VARIABLES 51
The limit expression in (4.63) shows that the sequence of random variables
λ
Yn ∼ Binomial n, , for n = 1, 2, 3, . . . ,
n
has a limiting distribution given by
λk −λ
pY (k) = e , for k = 0, 1, 2, 3, . . . (4.64)
k!
To see that the expression in (4.64) does indeed define a probability distribution
observe that
∞ ∞
X X λk −λ
pY (k) = e
k!
k=0 k=0
(4.65)
∞ k
X λ
= e−λ .
k!
k=0
Note that the expressions in (4.61) and (4.66) are well known realizations of
the exponential function x 7→ ex .
Definition 4.1.28 (Poisson Distribution). A discrete random variable, Y , which
can take on the values k = 0, 1, 2, . . ., is said to have a Poisson distribution with
parameter λ, if k
λ e−λ for k = 0, 1, 2, . . . ;
k!
pY (k) = (4.67)
0 elsewhere.
We write
Y ∼ Poisson(λ).
Example 4.1.29 (Expected Value of the Poisson Distribution). Let ∼ Poisson(λ).
Compute E(Y ).
Solution: Since Y takes on a countable number of values, the expected
value of Y is given by the series
∞
X
E(Y ) = mpY (m), (4.68)
m=0
52 CHAPTER 4. STOCHASTIC MODELS
where pY is given in (4.67). Thus, noting that the first term in the series in
(4.68) is zero, we obtain from (4.68) and (4.67) that
∞
X λm −λ
E(Y ) = m· e
m=1
m!
(4.69)
∞
Xλm
= e−λ .
m=1
(m − 1)!
∞
X λk · λ
= e−λ ,
k!
k=0
so that
∞
X λk
E(Y ) = λ e−λ . (4.70)
k!
k=0
Finally, use the series expansion for ex in (4.66) to obtain from (4.70) that
E(Y ) = λe−λ eλ = λ.
Thus, we have shown that
in other words, the expected value of a Poisson random variable with parameter
λ is λ. In Assignment #9 you’ll be asked to show that the variance, Var(Y ) of
of Y ∼ Poisson(λ) is also λ; where,
containing a virus population which is lethal to the bacteria which have not
developed resistance. Those bacteria which have mutated into resistant strains
will continue to replicate, while those that are sensitive to the virus will die.
After certain time, the resistant bacteria will develop visible colonies on the
plates. The number of these colonies will then correspond to the number of
resistant cells in each test tube at the time they were exposed to the virus. This
number corresponds to the number of bacteria in the colony that developed a
mutation which led to resistance. We denote this number by YN , where N is
the size of the colony after the nth division cycle. Assuming that the bacteria
may develop mutation to resistance after exposure to the virus, if N is very
large, according to the result in Section 4.1.5, the distribution of YN can be
approximated by a Poisson distribution with parameter λ = pN , where p is the
mutation rate and N is the size of the colony. It then follows that the probability
of no mutations occurring in one division cycle is
Pr(YN = 0) ≈ e−λ , (4.72)
according to (4.67). This probability can also be estimated experimentally as
Luria and nd Delbrück showed in their 1943 paper. In one of the experiments
described in that paper, out of 87 cultures of 2.4 × 108 bacteria, 29 showed not
resistant bacteria (i.e., none of the bacteria in the culture mutated to resistance
and therefore all perished after exposure to the virus). We therefore have that
29
Pr(YN = 0) ≈ .
87
Comparing this to the expression in Equation (4.72), we obtain that
29
e−λ ≈ ,
87
which can be solved for λ to obtain
29
λ ≈ − ln
87
or
λ ≈ 1.12.
The mutation rate, p, can then be estimated from λ = pN :
λ 1.12
p= ≈ ≈ 4.7 × 10−9 .
N 2.4 × 108
Example 4.1.30. Put four marbles in a bag. Two of the marbles are red and
the other two are blue. Pick two marbles at random and without replacement.
Labeling the marbles R1 , R2 , B1 and B2 , for the two red marbles and the two
blue marbles, respectively, we see that the sample space for the experiment (the
set of all possible outcomes of the experiment) can be written as
C = {R1 R2 , R1 B1 , R1 B2 , R2 B1 , R2 B2 , B1 B2 }.
The assumption of randomness in the picking of the two marbles implies that
each of the outcomes in C is equally likely; so that
1
Pr(c) = , for all c ∈ C. (4.74)
6
Let A denote the event that at least one of the marbles is red and B denote
the event that the two marbles have the same color; then,
A = {R1 R2 , R1 B1 , R1 B2 , R2 B1 , R2 B2 },
Proof: Assume that A and B are independent. Then, (4.73) holds true. Since
Pr(B) > 0, it follows from (4.73) and (4.79) that
Pr(A ∩ B) Pr(A) · Pr(B)
Pr(A | B) = = = Pr(A).
Pr(B) Pr(B)
Similarly, since Pr(A) > 0, it follows from (4.73) and (4.80) that
Pr(B | A) = Pr(B).
Then,
n
X
P (B) = P (Ek ) · P (B | Ek ) (4.83)
k=1
56 CHAPTER 4. STOCHASTIC MODELS
We assume that the the probability that service will be completed in the
short time interval [t, t + ∆t] is proportional to ∆t; say µ∆t, where µ > 0
is a proportionality constant. Then, the probability that service will not be
completed at t + ∆t is 1 − µ∆t. This situation is illustrated in the state diagram
pictured in Figure 4.1.1: The circles in the state diagram represent the possible
1 − µ∆t
?
# #
1 0
"!1"!
µ∆t
for ∆t small enough (see also the diagram in Figure 4.1.1), we therefore get
from (4.88) that
p(t + ∆t) ≈ p(t)(1 − µ∆t),
58 CHAPTER 4. STOCHASTIC MODELS
or
p(t + ∆t) − p(t) ≈ −µ∆t. (4.89)
Dividing both sides of (4.89) by ∆t 6= 0 and taking the limit as ∆t → 0 we
obtain that
p(t + ∆t) − p(t)
lim = −µp(t) (4.90)
∆t→0 ∆t
It follows from (4.90) that p is differentiable and satisfies the differential equa-
tion.
dp
= −µp(t). (4.91)
dt
The first order differential equation in (4.91) can be solved subject to the initial
condition p(0) = 1 to yield
Recall that T denotes the time it takes for service to be completed, or the
service time at the checkout counter. Thus, it is the case that
that is, T > t if and only if at time t the person is still at the checkout counter.
It follows from (4.93) and (4.87) that
To see that the function in (4.95) indeed defines a probability density func-
tion, compute
Z ∞ Z b
fT (t) dt = lim µe−µt dt, (4.96)
−∞ b→∞ 0
4.1. RANDOM VARIABLES 59
1.0
0.75
0.5
0.25
0.0
0 1 2 3 4 5
x
where Z b −µt b
µe−µt = −e 0
= 1 − e−µb (4.97)
0
since µ > 0.
Remark 4.1.37. It follows from (4.98) in Example 4.1.35 that the function
fT defined in (4.95) is a probability density function. We say that fT is the
probability density function of the random variable T .
Example 4.1.41 (Average Service Time). In the service time example, Exam-
ple 4.1.35, we showed that the time, T , that it takes for service to be completed
at a checkout counter has an exponential distribution with probability density
function given in (4.95),
(
µe−µt for t > 0,
fT (t) = (4.99)
0 otherwise,
where µ is a positive parameter. Note that for the expression in (4.99) to make
sense, the parameter µ has to have units of 1/time.
Observe that
Z ∞ Z ∞
|t|fT (t) dt = tµe−µt dt
−∞ 0
Z b
= lim t µe−µt dt
b→∞ 0
b
−µt 1 −µt
= lim −te − e
b→∞ µ 0
1 1
= lim − be−µb − e−µb
b→∞ µ µ
1
= ,
µ
where we have used integration by parts and L’Hospital’s rule. It then follows
that Z ∞
1
|t|fT (t) dt = < ∞
−∞ µ
and therefore the expected value of T exists and
Z ∞ Z ∞
1
E(T ) = tfT (t) dt = tµe−µt dt = .
−∞ 0 µ
4.2. RANDOM PROCESSES 61
Thus, the parameter µ is the reciprocal of the expected service time, or average
service time, at the checkout counter.
Example 4.1.42. Suppose the average service time, or mean service time,
at a checkout counter is 5 minutes. Compute the probability that a given person
will spend at least 6 minutes at the checkout counter.
Solution: By the result of Example 4.1.35, we assume that the service time,
T , has a probability density function given in (4.95) with µ = 1/5. We then
have that
Z ∞ Z ∞
1 −t/5
Pr(T > 6) = fT (t) dt = e dt = e−6/5 ≈ 0.30.
6 6 5
Thus, there is a 30% chance that a person will spend 6 minutes or more at the
checkout counter.
Definition 4.1.43 (Exponential Distribution). A continuous random variable,
X, is said to be exponentially distributed with parameter β > 0, written
X ∼ Exponential(β),
if it has a probability density function given by
1 −x/β
β e for x > 0,
fX (x) =
0 otherwise.
(i) P0 (0) = Pr[M (0) = 0] = 1; that is, initially no bacterium has mutated
into a strain of the particular type under study. It then follows that
(ii) The probability that any bacterium develops a mutation in a short time
interval [t, t + ∆t] depends only on ∆t and not on the number of mutant
bacteria at previous times.
(iii) The probability of a new mutation in the short interval [t, t + ∆t] is pro-
portional to ∆t; in symbols
The arrows in Figure 4.2.3 indicate the transition probabilities of going from
one state to the next, or those of remaining in the same state, in a short time
interval [t, t + ∆t]. For instance, if at time t there are no mutants in the colony
(i.e., the system is in state 0 at that time), then at time t + ∆t there might a
bacterium that has developed a mutation. The system would then go from state
0 to state 1 in the time interval [t, t + ∆t]; the probability of this occurrence
is approximately λ∆t by assumption (iii); this is indicated by the arrow in the
diagram that goes from state 0 to state 1. On the other hand, there might
not be a new mutation in the time interval [t, t + ∆t]; the probability of this
occurring is approximately 1 − λ∆t (why?), and this is shown by the arrow that
starts at state 0 and which winds back again to 0. Observe that assumption (iv)
is implicit in the state diagram in Figure 4.2.3 since the states can only increase
by 1 and not by 2 or more; thus, arrows from a given state either return to that
state or go to the next one.
4.2. RANDOM PROCESSES 63
The state diagram in Figure 4.2.3 can be used to estimate Pm (t + ∆t), given
that we know Pm (t), for very small values of ∆t. We start out with the case
m = 0 as follows:
since
Pr(no new mutations in [t, t+∆t] | M (t) = 0) = Pr(no new mutations in [t, t+∆t])
P0 (t + ∆t) − P0 (t)
≈ −λP0 (t). (4.103)
∆t
Next, let ∆t → 0 in (4.103) to conclude that P0 (t) is differentiable and
dP0
= −λP0 ; (4.104)
dt
that is, P0 (t) satisfies a first order differential equation. The differential equation
in (4.104) can be solved by separation of variables to yield
since, according to the state diagram in Figure 4.2.3, the system can get to
state 1 at t + ∆t via two routes: (i) from state 0 through a new mutation
which occurs with probability λ∆t, approximately, or (ii) from state 1 if no new
mutation occurs in the time interval [t, t + ∆t], and the approximate probability
64 CHAPTER 4. STOCHASTIC MODELS
of this occurrence is 1 −λ∆t. Here we have also used the law of total probability
in (4.83) and the independence assumption (ii).
Rearranging equation (4.107) and dividing by ∆t 6= 0, we obtain
P1 (t + ∆t) − P1 (t)
≈ −λP1 (t) + λP0 (t). (4.108)
∆t
Next, let ∆t → 0 in (4.108) to conclude that P1 is differentiable and satisfies
the differential equation
dP1
= −λP1 + λP0 (t)
dt
or, using (4.106),
dP1
= −λP1 + λe−λt . (4.109)
dt
The differential equation (4.109) can be solved as follows: Rewrite the equa-
tion as
dP1
+ λP1 = λe−λt (4.110)
dt
and multiply both sides of (4.110) by eλt to get
dP1
eλt + λeλt P1 = λ (4.111)
dt
Observe that, by the Product Rule,
d λt dP1
(e P1 ) = eλt + λeλt P1 ,
dt dt
and so the differential equation in (4.111) reduces to
d λt
(e P1 ) = λ. (4.112)
dt
The equation in (4.112) can be integrated to yield
eλt P1 = λt + C,
for t > 0.
Next, use the initial condition P1 (0) = 0 in (4.100), which follows from
assumption (i), to get that C = 0, and therefore
dP2
= −λP2 + λP1 (t)
dt
or, using (4.114),
dP2
= −λP2 + λ2 te−λt . (4.115)
dt
We can solve this differential equation as we solved (4.115), by first rearranging
and multiplying by eλt to get
dP2
eλt + λeλt P2 = λ2 t, (4.116)
dt
and then re–writing the left–hand side of (4.116), so that
d λt
(e P2 ) = λ2 t. (4.117)
dt
Next, integrate the equation in (4.117) and use the initial condition P2 (0) = 0
in (4.100) to get
(λt)2 −λt
P2 (t) = e , for t > 0. (4.118)
2
One can go through the same procedure leading to (4.118) to obtain the
formula
(λt)3 −λt
P3 (t) = e
3!
for P3 (t), and this suggests the general formula for Pm (t), m = 0, 1, 2, . . ., to be
(λt)m −λt
Pm (t) = e , for t > 0. (4.119)
m!
We will establish the formula in (4.119) by induction on m. Observe that we
have already established the basic case m = 0 in (4.106). Next, for the inductive
step, assume that the formula (4.119) holds for m, and we seek to show that it
also holds for m + 1. Using the state diagram 4.2.3 we see that
d
(Pm+1 ) = −λPm+1 + λPm (t)
dt
or, using the inductive hypothesis (4.119),
d λm+1 tm −λt
(Pm+1 ) = −λPm+1 + e . (4.120)
dt m!
66 CHAPTER 4. STOCHASTIC MODELS
We can solve the differential equation in (4.120) as we solved (4.115); that is,
first rearrange the equation and multiply by eλt to get
d λm+1 tm
eλt (Pm+1 ) + λeλt Pm+1 = ; (4.121)
dt m!
then, re–write the left–hand side of the equation in (4.121) to get
d λt λm+1 tm
(e Pm+1 ) = . (4.122)
dt m!
Integrating (4.122) and using the initial condition Pm+1 (0) = 0 in (4.100), we
obtain
(λt)m+1 −λt
Pm+1 (t) = e
(m + 1)!
for all t ≥ 0, since (m+1)! = (m+1)m!. This establishes the formula (4.119) for
the case m+1, and therefore formula (4.119) is now proved for all m = 0, 1, 2, . . .
by induction on m.
Note that the formula in (4.119),
(λt)m −λt
Pm (t) = e , for m = 0, 1, 2, 3, . . . and t > 0, (4.123)
m!
is the probability distribution of a Poisson(λt) random variable. We have there-
fore demonstrated that assumptions (i)–(iv) imply that, for each t > 0, M (t) is
a Poisson random variable with parameter λt,
We say that M (t) is a Poisson random process. This particular random process
is characterized by the assumptions (i)–(iv) that we made on M (t).
In general, a random process is a collection, {X(t) | t ∈ I}, of random
variables, X(t), for t in some indexing set I. If I is countable, for instance,
I = N or I = Z, the random process {X(t) | t ∈ I} is called a discrete–time
random process. If I is some interval of real numbers, then {X(t) | t ∈ I} is
called a continuous–time random process. For the case of the Poisson random
process defined in (4.123), I = [0, ∞). Therefore, the Poisson random process
is a continuous–time random process.
Given a random process, {X(t) | t ∈ I}, the mean of the random process is
given by
E[X(t)], for t ∈ I.
It follows from (4.124), or (4.123), and the calculations in Section 4.1.5 that
Thus, assumptions (i)–(iv) imply that the expected number of mutations in the
interval [0, t] is proportional to the length of the time interval. The constant of
proportionality, λ, represents the average number of mutations per unit time.
Bibliography
67