CK1 Booklet 1 PDF
CK1 Booklet 1 PDF
Subject CS2
Revision Notes
For the 2019 exams
Markov models
Booklet 2
Covering
CONTENTS
Contents Page
Copyright agreement
Legal action will be taken if these terms are infringed. In addition, we may
seek to take disciplinary action through the profession or through your
employer.
These conditions remain in force after you have finished using the course.
These chapter numbers refer to the 2019 edition of the ActEd course notes.
The numbering of the syllabus items is the same as that used by the Institute
and Faculty of Actuaries.
OVERVIEW
This booklet covers Syllabus objectives 3.2, 4.1.8 and 4.3, which relate to
Markov chains, the two-state Markov model and the Poisson model.
Markov chain models are stochastic processes with a discrete time set and a
discrete state space (usually a series of integers or labels). It is difficult to
use time-inhomogeneous models effectively, so we concentrate mainly on
time-homogeneous models.
Simple methods for the simulation of a Markov chain model complete the
discussion of this topic.
Sometimes the model you are given initially may not actually be a Markov
model and you have to modify it so that it is Markov.
The two-state Markov model and the Poisson model are parametric mortality
models. We can use these models to estimate m x based on the results of a
mortality investigation.
In the exam, you could be asked to write down the likelihood function
associated with a model and then derive the maximum likelihood estimate of
the parameter to be estimated.
CORE READING
All of the Core Reading for the topics covered in this booklet is contained in
this section.
The text given in Arial Bold Italic font is additional Core Reading that is not
directly related to the topic being discussed.
____________
3 The Markov property (2.1) has the following interpretation: given the
present state of the process X m i , the additional knowledge of the
past is irrelevant for the calculation of the probability distribution of
future values of the process.
____________
4 The conditional probabilities on the right-hand side of (2.1) are the key
objects for the description of a Markov chain; we call them transition
probabilities, and we denote them by:
P X n jX m i pi(m ,n )
j
____________
m ,l p l ,n
pi(m
j
,n )
pi k kj
k S
Proof
[Students should understand this proof, but they will not be expected
to reproduce it in the examination.]
This is based on the Markov property (2.1) and on the law of total
probability in its conditional form.
Ak , Ak A j , k j
k 1
P BC P BC , Ak P AkC
k 1
Thus:
P X n jX l k P X l kX m i
k S
P X 0 i0 , X 1 i1,..., X n in qi0 pi
0,1 p(1,2) ... p(n 1,n )
i
0 1 i ,i
1 2 i i
n 1, n
l
P X l m jX m i pij (2.3)
nm l m p n l
pij pik kj for m < l < n
k S
____________
Pij pij
____________
l
pij P l ij
____________
l
pij 1 for all i
j S
____________
Models
3/4 3/4
3/4
State 0 State 1 State 2
0% 25% 50%
discount discount discount
1/4
1/4
1/4
14 3
4 0
P 14 0 3
4
0 1 3
4 4
3
p0,2 P 3
1,3
9 16
____________
P01(t) P12(t)
P00(t)
P10(t) P21(t)
and:
17 Modify the previous model as follows: there are now four levels of
discount:
0 : no discount
1 : 25% discount
2 : 40% discount
3 : 60% discount
The rules for moving up the discount scale are as before, but in the
case of a claim during the current year, the discount status moves
down one or two steps (if this is possible) according to whether or not
the previous year was claim-free.
P X n 1 0X n 2, X n 1 1 0
whereas:
P X n 1 0X n 2, X n 1 3 0 .
____________
3/4
State 0 State 1 State 2+ State 3
0% 25% 40% 60%
discount discount discount 1/4 discount
1/4
1/4 1/4
3/4
State 2-
40%
discount
1/4
____________
14 3
4 0 0 0
1 3
4 0 4 0 0
P 0 1
4 0 0 3
4
14 0 0 0 3
4
0 0 0 1 3
4 4
____________
3
p1,3 P 3
2,5
27 64
____________
Random walks
p p p p p p
-2 -1 0 1 2
È ˘
Í ˙
Í ˙
Í 1 - p 0 p ˙
Í ˙
Í 1 - p 0 p ˙
P = Í ˙
Í ˙
Í 1- p 0 p ˙
Í 1- p 0 p ˙˙
Í
Í ˙
Í ˙
Î ˚
____________
n u n u
p 1 p if 0 n j i 2n and n j i is even
pij(n ) u
0 otherwise
____________
n n
pi j pi r j r
____________
Reflecting boundary:
P X n 1 1X n 0 1
Absorbing boundary:
P X n 1 0X n 0 1
Mixed boundary:
P X 0X n 0 ,
n 1
P X n 1 1X n 0 1
____________
1– p p p p p p p
0 1 2 j –1 j j +1 b–2 b–1 b
1–p 1–p 1–p 1–p 1–p 1–p 1–p 1–
1 2 3 b-2 b -1 b
È a 1- a ˘
Í ˙
Í1 - p 0 p ˙
Í 1- p 0 p ˙
Í ˙
Í 1 - p 0 p ˙
Í
P= ˙
Í ˙
Í 1- p 0 p ˙
Í 1- p 0 p ˙˙
Í
Í 1- p 0 p˙
Í ˙
Î 1- b b˚
____________
f ( y 1 + y 2 + ... + y n )
P ÈYn +1 = 1 ΩY1 = y 1, Y2 = y 2 , ... , Yn = y n ˘ =
Î ˚ g (n )
Of course:
f ( y 1 + y 2 + ... + y n )
P ÈYn +1 = 0ΩY1 = y 1, Y2 = y 2 , ... , Yn = y n ˘ = 1 -
Î ˚ g (n )
____________
31 The dependence on the past record means that Y1, Y2 , ... , Yn , ... does
not have the Markov property (it depends on all previous values of Y j ).
____________
n
Xn Yj
j 1
n
Since Yj X n , the condition Y1 x1, Y2 x 2 x1, ... , Yn x n x n 1
j 1
j i pij (2.4)
iS
j 0
j 1
jS
P X 1 j P X 1 jX 0 i P X 0 i
pij i j
i S i S
36 A Markov chain with a finite state space has at least one stationary
probability distribution.
____________
0 1
4 0 1
4 1
1
4 2
1 3 4 0 1
4 2
2 3 4 1 (2.5)
2 1
4 3
3 3 4 2 3 4 2 3 4 3
____________
37 This linear system is not linearly independent since adding up all the
equations results in an identity (this is a general feature of equations
P due to the property pij 1 ). Because of this we can
jS
discard any one of the equations, say the last one.
Step 1: Discard one of the equations. Here the first or the last one are
obvious choices; delete the final one say.
3 0 2 1 (a)
3 0 2 4 1 (b)
2 3 4 1 (c)
4 2 3 0 (d)
p 2+ = 3 4 p 1
p0 =
p1
3 (4 - 3 4) = 13 12 p 1
p 2 - = p 1 ( -1 + 13 4 ) = 9 4 p 1
p 3 = 9p 1
p1
Âp j = 12 (13 + 12 + 9 + 27 + 108) = 169 12 p 1 = 1
j
p 1 = 12 169
Step 6: Combine the results of the last two steps to obtain the
solution.
____________
____________
n can exist.
It is only for aperiodic states that lim pii
n
One can check using the transition graphs that, in the no claims
discount models on pages 10 and page 12, all states are aperiodic
whereas in the model of the simple random walk without boundaries on
page 14, all states have period 2. Finally in the simple random walk
model with boundaries on page 16 all states are aperiodic unless both
and are either 0 or 1.
____________
44 If a Markov chain is irreducible all its states have the same period (or
all are aperiodic).
____________
45
n be the n -step transition probability of an irreducible aperiodic
Let pij
Markov chain on a finite state space. Then for every i and j :
lim pij
n
j
n
Note how the above limit is independent of the starting state i . This
proof is beyond the syllabus.
46 The first thing to fix when setting up a Markov model is the state space.
As shown by the example of a no claims discount model on page 12,
the state space which first springs to mind may not be the most
suitable and may need some modification before a Markov model can
be fitted.
Once the state space is determined, however, the Markov model must
be fitted to the data by estimating the transition probabilities pij .
____________
nij
Then the best estimate of pij is pˆ ij = .
ni
____________
The next step is to ensure that the fit of the model to the data is
adequate, or in other words to check that the Markov property seems
to hold.
____________
nijk nij pˆ jk
2
2
i j k nij pˆ jk
____________
pi ,i +1 = q , pi ,i -1 = f , pi ,i = 1 - q - f
Simulation
InitialState = c(1,0,0)
After3Months = InitialState * (Employment *
Employment * Employment)
After6Months = InitialState * (Employment^6)
x
a = alive d = dead
____________
The probability that a life alive at a given age will be dead at any
subsequent age is governed by the age-dependent transition
intensity m x +t (t ≥ 0) , in a way made precise by Assumption 2 below.
____________
56 Assumption 1
The probabilities that a life at any given age will be found in either state
at any subsequent age depend only on the ages involved and on the
state currently occupied. This is the Markov assumption.
____________
(b) specify a model which took them into account; in other words,
treat the problem as one of regression.
____________
58 Assumption 2
dt q x +t = m x +t dt + o(dt ) (t ≥ 0 )
____________
60 Assumption 3
+ t q x ¥ P [Alive at x + t + dt | Dead at x + t ]
= ( t px ¥ dt px +t ) + ( t q x ¥ 0)
= t px ¥ (1 - m x +t dt + o(dt ))
Therefore:
∂ t + dt p x- t px
p = lim
∂ t t x dt Æ0 + dt
o(dt )
= - t px m x +t + lim
dt Æ0 + dt
= - t p x m x +t (3.1)
Ê t ˆ
So t px = exp Á - Ú m x + s ds ˜ as before.
Ë 0 ¯
____________
The important point is that it has been derived here strictly from the
assumptions of the two-state model, and that the method is easily
extended to models with more states. In the Markov framework, (3.1) is
an example of the Kolmogorov forward equations. These are
discussed in detail in Booklet 3.
____________
Di = 0 ¤ Ti = bi , Di = 1 ¤ ai < Ti < bi
____________
68 It will often be useful to work with the time spent under observation, so
define Vi = Ti - ai . Vi is called the waiting time. It has a mixed
distribution, with a probability mass at the point bi - ai .
____________
69 The pair (Di ,Vi ) comprise a statistic, meaning that the outcome of our
observation is a sample (d i , v i ) drawn from the distribution of (Di ,Vi ) .
ÏÔ bi - ai px + ai (d i = 0)
fi (d i , v i ) = Ì
ÔÓ v i px + ai m x + ai +v i (d i = 1)
Ï Ê bi - ai ˆ
Ôexp Á - Ú m x + a +t dt ˜ (d i = 0)
Ô ÁË i ˜¯
Ô 0
=Ì
Ô Ê vi ˆ
Ô exp Á - Ú m x + ai +t dt ˜ m x + ai +v i (d i = 1)
ÔÓ ÁË 0 ˜¯
Ê vi ˆ
d
= exp Á - Ú m x + ai +t dt ˜ m x i+ a +v
ÁË 0 ˜¯ i i
fi (d i , v i ) = e - mv i m d i
____________
70 The joint probability function of all the (Di ,Vi ) , by independence, is:
i =N
’ e - mv i m d i = e - m (v1 + ... + v N ) m d i + ... + d N = e - m v m d
i =1
N N
where d = Â d i and v = Â v i . In other words define random variables
i =1 i =1
D and V to be the total number of deaths and the total waiting time,
respectively, and the joint probability function of all the (Di ,Vi ) can be
simply expressed in terms of D and V .
____________
L ( m ; d , v ) = e - mv m d
mˆ = d / v
m = D / V
E ÎÈDi - m Vi ˚˘ = 0 (3.2)
Note that the first of these can also be written as E ÈÎDi ˘˚ = m E ÈÎVi ˘˚ .
____________
73 In the case where the {ai } and {bi } are known constants, this follows
from integrating/summing the probability function of (Di , Vi ) over all
possible events to obtain:
bi - ai
Ú e - m v i m dv i + e - m (bi - ai ) = 1
0
1 1 N
N
(D - m V ) = N Â (Di - mVi )
i =1
N Ê D mV ˆ
lim ( m - m ) = lim ÁË N - N ˜¯
N Æ• N Æ• V
Ê m ˆ
m ~ Normal Á m ,
Ë E [V ] ˜¯
____________
77 That is:
c
e - m E x ( m E xc )d
P[D = d ] =
d!
____________
D
m =
E xc
m
E [ m ] = m and var[ m ] =
E xc
m
80 Under the two-state model, E [ m ] = m and var[ m ] = , but the true
E[V ]
values of m and E [V ] are unknown and must be estimated from the
data as m̂ and E xc respectively.
qˆ x = 1 - e - m
ˆ
____________
Comment on application
Ê t ˆ
t px = exp Á - Ú m x + s ds ˜
Ë 0 ¯
This section contains all the relevant exam questions from 2008 to 2017 that
are related to the topics covered in this booklet.
Solutions are given after the questions. These give enough information for
you to check your answer, including working, and also show you what an
outline examination answer should look like. Further information may be
available in the Examiners’ Report, ASET or Course Notes. (ASET can be
ordered from ActEd.)
We first provide you with a cross-reference grid that indicates the main
subject areas of each exam question. You can use this, if you wish, to
select the questions that relate just to those aspects of the topic that you
may be particularly interested in reviewing.
Alternatively, you can choose to ignore the grid, and attempt each question
without having any clues as to its content.
Cross-reference grid
Question attempted
Time-homog / time-
Two-state Markov
Transition matrix
Markov property
Transition graph
Poisson model
stationary dist
Probabilities /
Irreducibility
State space
Periodicity
Question
inhomog
model
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
Page 40
Markov property
State space
Transition graph
Transition matrix
Irreducibility
Periodicity
Probabilities /
stationary dist
Batch3
Time-homog /
Time-inhomog
Two-state Markov
model
Poisson model
Question attempted
In a certain small country all listed companies are required to have their
accounts audited on an annual basis by one of the three authorised audit
firms (A, B and C). The terms of engagement of each of the audit firms
require that a minimum of two annual audits must be conducted by the newly
appointed firm. Whenever a company is able to choose to change auditors,
the likelihood that it will retain its auditors for a further year is (80%, 70%,
90%) where the current auditor is (A,B,C) respectively. If changing auditors
a company is equally likely to choose either of the alternative firms.
(i) A company has just changed auditors to firm A. Calculate the expected
number of audits which will be undertaken before the company changes
auditors again. [2]
(ii) Formulate a Markov chain which can be used to model the audit firm
used by a company, specifying:
(iii) Calculate the expected proportion of companies using each audit firm in
the long term. [5]
[Total 11]
Level 1: 0% discount
Level 2: 25% discount
Level 3: 40% discount
Level 4: 60% discount
(i) Write down the transition matrix of this No-Claims Discount process. [1]
n
Consider the random variable defined by X n = Â Yi with each Yi mutually
i =1
independent with probability:
(i) Write down the state space and transition graph of the sequence X n .
[2]
(a) is aperiodic
(b) is reducible
(iv) Derive expressions for the m-step transition probabilities pij( m ) . [3]
(v) Show how the one-step transition probabilities would alter if X n was
restricted to non-negative numbers by introducing:
(vi) For each of the examples in part (v), explain whether the transition
probabilities pij( m ) would increase, decrease or stay the same.
(Calculation of the transition probabilities is not required.) [3]
[Total 15]
Ê1 - a a ˆ
ÁË b 1 - b ˜¯
(ii) Explain the range of values that a and b can take which result in this
being a valid Markov chain which is:
(a) irreducible
The rules governing a policyholder’s discount level, based upon the number
of claims made in the previous year, are as follows:
Following a year with no claims, the policyholder moves up one discount
level, or remains at the 60% level.
Following a year with one claim, the policyholder moves down one
discount level, or remains at the 0% level.
Following a year with two or more claims, the policyholder moves down
two discount levels (subject to a limit of the 0% discount level).
(i) Determine the transition matrix for the no claims discount system. [3]
The following data shows the number of the insurer’s 130,200 policyholders
in the portfolio classified by the number of claims each policyholder made in
the last year. This information was used to estimate the mean of 0.30.
No claims 96,632
One claim 28,648
Two claims 4,400
Three claims 476
Four claims 36
Five claims 8
(iv) Test the goodness of fit of these data to a Poisson distribution with
mean 0.30. [5]
(v) Comment on the implications of your conclusion in (iv) for the average
level of discount applied. [2]
[Total 16]
(ii) Prove that if the process has independent increments it satisfies the
Markov property. [3]
(iii) (a) Describe the difference between a Markov chain and a Markov
jump process.
An actuarial student can see the office lift (elevator) from his desk. The lift
has an indicator which displays on which of the office’s five floors it is at any
point in time. For light relief the student decides to construct a model to
predict the movements of the lift.
(a) irreducible
A firm rents cars and operates from three locations – the Airport, the Beach
and the City. Customers may return vehicles to any of the three locations.
The company estimates that the probability of a car being returned to each
location is as follows:
Car returned to
Car hired from Airport Beach City
(iv) Sketch, using your answers to parts (i) and (ii), a graph showing the
probability that a car currently located at the Airport is subsequently at
the Airport, Beach or City against the number of times the car has been
rented. [3]
[Total 10]
A Markov Chain with state space {A, B, C} has the following properties:
it is irreducible
it is periodic
the probability of moving from A to B equals the probability of moving
from A to C.
(i) Show that these properties uniquely define the process. [4]
(i) (a) Explain how this scheme can be modelled as a Markov chain.
(b) Explain why there must be a unique stationary distribution for the
proportion of members in each class. [3]
The airline’s research has shown that in any given year, 40% of members
book no flights, 40% book exactly one flight, and 20% book two or more
flights.
The cost of running the scheme per member per year is as follows:
Ordinary members £0
Bronze members £10
Silver members £20
Gold members £30
The airline makes a profit of £10 per passenger for every flight before taking
into account costs associated with the frequent flyer scheme.
(iv) Assess whether the airline makes a profit on the members of the
scheme. [4]
[Total 13]
A pet shop has four glass tanks in which snakes for sale are held. The shop
can stock at most four snakes at any one time because:
if more than one snake were held in the same tank, the snakes would
attempt to eat each other and
having snakes loose in the shop would not be popular with the
neighbours.
The number of snakes sold by the shop each day is a random variable with
the following distribution:
If the shop has no snakes in stock at the end of a day, the owner contacts
his snake supplier to order four more snakes. The snakes are delivered the
following morning before the shop opens. The snake supplier makes a
charge of C for the delivery.
(i) Write down the transition matrix for the number of snakes in stock when
the shop opens in a morning, given the number in stock when the shop
opened the previous day. [2]
(ii) Calculate the stationary distribution for the number of snakes in stock
when the shop opens, using your transition matrix in part (i). [4]
(iii) Calculate the expected long term average number of restocking orders
placed by the shop owner per trading day. [2]
(iv) Calculate the expected long term number of sales lost per trading day.
[2]
The owner is unhappy about losing these sales as there is a profit on each
sale of P . He therefore considers changing his restocking approach to
place an order before he has run out of snakes. The charge for the delivery
remains at C irrespective of how many snakes are delivered.
(v) Evaluate the expected number of restocking orders, and number of lost
sales per trading day, if the owner decides to restock if there are fewer
than two snakes remaining in stock at the end of the day. [5]
(vi) Explain why restocking when two or more snakes remain in stock
cannot optimise the shop’s profits. [2]
The pet shop owner wishes to maximise the profit he makes on snakes.
(vii) Derive a condition in terms of C and P under which the owner should
change from only restocking where there are no snakes in stock, to
restocking when there are fewer than two snakes in stock. [2]
[Total 19]
Children at a school are given weekly grade sheets, in which their effort is
graded in four levels: 1 ‘Poor’, 2 ‘Satisfactory’, 3 ‘Good’ and 4 ‘Excellent’.
Subject to a maximum level of Excellent and a minimum level of Poor,
between each week and the next, a child has:
a 20 per cent chance of moving up one level
a 20 per cent chance of moving down one level
a 10 per cent chance of moving up two levels
a 10 per cent chance of moving down two levels.
Children are graded on Friday afternoon in each week. On Friday of the first
week of the school year, as there is little evidence on which to base an
assessment, all children are graded ‘Satisfactory’.
(ii) Calculate the probability distribution of the process after the grading on
Friday of the third week of the school year. [3]
[Total 5]
Farmer Giles makes hay each year and he makes far more than he could
possibly store and use himself, but he does not always sell it all. He has
decided to offer incentives for people to buy large quantities so it does not sit
in his field deteriorating. He has devised the following “discount” scheme.
He has a Base price B of £8 per bale. Then he has three levels of discount:
Good price G is a 10% discount, Loyalty price L is a 20% discount and
Super price S is a 25% discount on the Base price.
Customers who increase their order compared with last year move to
one higher discount level, or remain at level S .
Customers who maintain their order from last year stay at the same
discount level.
Customers who reduce their order from last year drop one level of
discount or remain at level B provided that they maintained or
increased their order the previous year.
Customers who reduce their order from last year drop two levels of
discount if they also reduced their order last year, subject to remaining
at the lowest level B .
(i) Explain why a process with the state space {B, G, L, S} does not display
the Markov property. [2]
(ii) (a) Define any additional state(s) required to model the system with the
Markov property.
Farmer Giles thinks that each year customers have a 60% likelihood of
increasing their order and a 30% likelihood of reducing it, irrespective of the
discount level they are currently on.
(iii) (a) Write down the transition matrix for the Markov process.
(c) Hence calculate the long-run average price he will get for each bale
of hay. [8]
(iv) Calculate the probability that a customer who is currently paying the
Loyalty price L will be paying L in two years’ time. [3]
(v) Suggest reasons why the assumptions Farmer Giles has made about
his customers’ behaviour may not be valid. [3]
[Total 19]
The diagrams below show three Markov chains, where arrows indicate a
non-zero transition probability:
A State 1 State 2
State 1 State 2
State 3 State 4
An actuary walks from his house to the office each morning, and walks back
again each evening. He owns two umbrellas. If it is raining at the time he
sets off, and one or both of his umbrellas is available, he takes an umbrella
with him. However if it is not raining at the time he sets off he always forgets
to take an umbrella.
Assume that the probability of it raining when he sets off on any particular
journey is a constant p , independent of other journeys.
(i) Explain why the transition graph for this process is given by:
1 p
Zero Two One 1–p
1–p p
[3]
(ii) Derive the transition matrix for the number of umbrellas at the actuary’s
house before he leaves each morning, based on the number before he
leaves the previous morning. [3]
(iii) Calculate the stationary distribution for the Markov Chain. [3]
(iv) Calculate the long run proportion of journeys (to or from the office) on
which the actuary sets out in the rain without an umbrella. [2]
The actuary considers that the weather at the start of a journey, rather than
being independent of past history, depends upon the weather at the start of
the previous journey. He believes that if it was raining at the start of a
journey the probability of it raining at the start of the next journey is r
(0 < r < 1) , and if it was not raining at the start of a journey the probability of
it raining at the start of the next journey is s (0 < s < 1, r π s ) .
(v) Write down the transition matrix for the Markov Chain for the weather. [1]
(vi) Explain why the process with three states {0,1, 2} , being the number of
his umbrellas at the actuary’s current location, would no longer satisfy
the Markov property. [2]
(vii) Describe the additional state(s) needed for the Markov property to be
satisfied, and draw a transition diagram for the expanded system. [4]
[Total 18]
The series Yi records, for each time period i , whether a car driver is
accident-free during that period ( Yi = 0 ) or has at least one accident
( Yi = 1 ).
i
Define X i = Â Yj with state space {0,1,2,} .
j =1
1Ê X ˆ 1
P (Yn +1 = 1) = 1 + n ˜ for n ≥ 1 , with P (Y1 = 1) = .
4 ÁË n ¯ 2
(i) Demonstrate that the series X i satisfies the Markov property, whilst Yi
does not. [2]
(a) irreducible
(iii) Draw the transition graph for X i covering all transitions which could
occur in the first three time periods, including the transition probabilities.
[4]
(iv) Calculate the probability that the driver has accidents during exactly two
of the first three time periods. [2]
(ii) Identify the minimum number of states under which the payments under
the scheme can be modelled using a time-homogeneous Markov chain,
specifying these states. [2]
(vi) Calculate the contributions required if, instead, sick pay continued at
100% of salary indefinitely. [2]
(vii) Comment on the benefit to the scheme of the reduction in sick pay to
50% from the third month. [2]
[Total 15]
A no claims discount system operates with three levels of discount, 0%, 15%
and 40%. If a policyholder makes no claim during the year he moves up a
level of discount (or remains at the maximum level). If he makes one claim
during the year he moves down one level of discount (or remains at the
minimum level) and if he makes two or more claims he moves down to, or
remains at, the minimum level.
The probability for each policyholder of making two or more claims in a year
is 25% of the probability of making only one claim.
The long-term probability of being at the 15% level is the same as the
long-term probability of being at the 40% level.
(i) Derive the probability of a policyholder making only one claim in a given
year. [4]
(ii) Determine the probability that a policyholder at the 0% level this year
will be at the 40% level after three years. [2]
(iii) Estimate the probability that a policyholder at the 0% level this year will
be at the 40% level after 20 years, without calculating the associated
transition matrix. [3]
[Total 9]
The stadium has an arrangement with the Floodwatch repair company who
are brought in the morning after a floodlight breakdown and charge $1,000
per day. There is a 60% chance they are able to repair the floodlights such
that the evening game can take place and be completed without needing to
be abandoned. If they are still broken the repair company is used (and paid)
again each day until the lights are fixed, with the same 60% chance of fixing
the lights each day.
(ii) Write down the transition matrix for the process which describes
whether the floodlights are working or not. [1]
(iii) Derive the long run proportion of games which have to be abandoned.
[3]
(iv) Calculate the maximum amount the stadium should be prepared to pay
Light Fantastic to improve profitability. [4]
[Total 10]
(iii) (a) State how many states are required to model this as a Markov
chain.
(iv) Calculate the proportion of policyholders in the long run who are at the
25% level. [6]
The two football teams in a particular city are called United and City and
there is intense rivalry between them. A researcher has collected the
following history on the results of the last 20 matches between the teams
from the earliest to the most recent, where:
UCCDDUCDCUUDUDCCUDCC
The researcher has assumed that the probability of each result for the next
match depends only on the most recent result. He therefore decides to fit a
Markov chain to this data.
(i) Estimate the transition probabilities for the Markov chain. [3]
(ii) Estimate the probability that United will win at least two of the next three
matches against City. [3]
[Total 6]
(i) Write down the transition matrix of the Markov chain with state space
{0%, 25%, 50%, 60%}. [2]
(a) irreducible
(iv) Explain, without doing any further calculations, how the answers to parts
(ii) and (iii) would change as a result of introducing the ‘protected’
No Claims Discount scheme. [3]
[Total 11]
(i) Derive the maximum likelihood estimator under the Poisson model of
the average rate at which events occur, m , in a population where the
exposed to risk for each person i is E . [4]
A university runs a bus service between its teaching campus and its student
halls of residence. Traffic conditions mean that the arrival of buses at the
bus stop on the teaching campus can be considered to follow a Poisson
process.
Student Time arrived Time left (if left before Time next bus
next bus arrived) arrived
1 4:00 pm 4:05 pm
2 4:10 pm 4:35 pm
3 4:20 pm 4:30 pm
4 4:30 pm 4:35 pm
5 4:40 pm 4:50 pm
6 4:45 pm 4:50 pm
7 4:55 pm 5:05 pm
8 5:00 pm 5:20 pm
9 5:10 pm 5:40 pm
10 5:10 pm 6:10 pm
(ii) Calculate the maximum likelihood estimate of the hourly rate at which
buses arrive at the bus stop, using the Poisson estimator, and assuming
that only one bus arrived at any given time. [3]
(iii) Comment on the use of the Poisson model for this investigation. [3]
[Total 10]
The probabilities of the heating element being in each condition at the end of
a cycle, based on the condition at the start of the cycle are as follows:
START END
Excellent Good Poor Failed
Excellent 0.5 0.2 0.2 0.1
Good 0.5 0.3 0.2
Poor 0.5 0.5
Failed 1
(i) Write down the name of the stochastic process which describes the
condition of a single heating element over time. [1]
(ii) Explain whether the process describing the condition of a single heating
element is:
(a) irreducible
(iii) Derive the probability that the condition of a single heating element is
assessed as being in Poor condition at the inspection after two cycles, if
the heating element is currently in Excellent condition. [2]
If the heating element fails during the firing cycle, the entire batch of tiles in
the kiln is wasted at a cost of £1,000. Additionally, a new heating element
needs to be installed at a cost of £50, which will, of course, be in Excellent
condition.
(iv) Write down the transition matrix for the condition of the heating element
in the kiln at the start of each cycle, allowing for replacement of failed
heating elements. [2]
(v) Calculate the long-term probabilities for the condition of the heating
element in the kiln at the start of a cycle. [4]
(vi) Calculate the expected annual cost incurred due to failures of heating
elements. [2]
The company is concerned about the cost of ruined tiles and decides to
change its policy to replace the heating element if it is rated as in Poor
condition.
A sports league has two divisions {1,2} with Division 1 being the higher.
Each season the bottom team in Division 1 is relegated to Division 2, and
the top team in Division 2 is promoted to Division 1.
(i) Write down the minimum number of states required to model this as a
Markov chain. [1]
(iii) Write down the transition matrix for the Markov chain. [2]
(a) irreducible
0: no discount
1: 15% discount
2: 25% discount
3: 40% discount
A policyholder who does not make a claim in the year moves up one level of
discount the following year (or stays at the maximum level).
A policyholder who makes one or more claims in a year moves down one
level of discount if they did not claim in the previous year (or remains at the
lowest level) but if they made at least one claim in the previous year they
move down two levels of discount (subject to not going below the lowest
level).
(i) (a) Explain how many states are required to model this as a Markov
chain.
The probability p of making at least one claim in any year is constant and
independent of whether a claim was made in the previous year.
(ii) Calculate the proportion of policyholders who are at the 25% discount
level in the long run given that the proportion at the 40% level is nine
times that at the 15% level. [6]
(iii) (a) Explain how the state space of the process would change if the
probability of making a claim in any one year depended upon
whether a claim was made in the previous year.
(b) Write down the transition matrix for this new process. [4]
[Total 13]
A simplified model of the internet consists of the following websites with links
between the websites as shown in the diagram below.
N(ile) B(anana)
C(heep) H(andbook)
(ii) Calculate the transition matrix for the Markov chain representing which
website the internet user is on. [2]
(iii) Calculate, of the total number of visits, what proportion are made to
each website in the long term. [4]
[Total 8]
Yolanda has just bought her policy from Company C for the first time.
(i) Calculate the probability that Yolanda will be covered by Company C for
at least five years before she changes provider. [2]
(ii) Calculate the probability that Company A does NOT cover Zachary’s
home at the time of the fire. [2]
(iv) Determine the transition matrix which will apply after the takeover if
Company A’s assumption about homeowners’ behaviour is correct. [2]
The diagrams below show three Markov chains, where arrows indicate a
non-zero transition probability.
A Markov Chain 1
State 1
State 2 State 3
B Markov Chain 2
State 1 State 2
State 3 State 4
C Markov Chain 3
State 1 State 2
An individual’s marginal tax rate depends upon his or her total income during
a calendar year and may be 0% (that is, he or she is a non-taxpayer), 20%
or 40%.
The movement in the individual’s marginal tax rate from year to year is
believed to follow a Markov chain with a transition matrix as follows:
0% Ê1 - b - b 2 b b2 ˆ
Á ˜
20% Á b 1 - 3b 2b ˜
40% Á 2 2˜
Ë b b 1- b - b ¯
(i) Draw the transition diagram of the process, including the transition
rates. [2]
(ii) Determine the range of values of b for which this is a valid transition
matrix. [3]
(a) irreducible
(b) periodic
(iv) Calculate the long term proportion of taxpayers at each marginal rate. [4]
(v) Calculate the probabilities that Lucy’s marginal tax rate in 2013 is:
(a) 0%
(b) 20%
The profession would like the city to increase the size of the rack. The city
has said it will do so if the profession can demonstrate that, in the long run,
the rack is full or empty for more than 35 per cent of the working day. The
profession commissions a study to monitor the rack every 10 minutes during
the working day.
(i) Give the transition matrix for the number of bicycles in the rack. [2]
(ii) Determine whether the city will increase the size of the rack. [6]
(iii) Comment on whether an increase in the size of the rack will reduce the
proportion of time for which the rack is empty or full. [2]
[Total 10]
The previous foreign correspondent tells him ‘If you want to know how many
flights you are likely to take, I estimate my movements have been like this’
and she drew this diagram showing the transition probabilities:
On his first day in the job the new foreign correspondent will be in Atlantis.
(ii) Calculate the probability that the foreign correspondent will be in each
country in his third day in the job. [2]
The previous correspondent also reported that Beachy must be the most
interesting of the islands in terms of news because she spent 41.9% of her
time there compared with 32.6% on Atlantis and 25.6% on Coral.
(iii) Sketch a graph showing the probability that the journalist is in each
country over time, using the information above. [3]
(iv) Calculate the proportion of days on which the foreign correspondent will
take a flight. [1]
The first time the foreign correspondent visits each of the countries he takes
a photograph to mark the occasion.
(v) Identify a suitable state space for modelling as a Markov chain which
countries he has visited so far. [2]
(vi) Draw a transition diagram for the possible transitions between these
states. [3]
[Total 12]
The following is a partially completed transition matrix for this Markov Chain:
A Ê 0.2 - - ˆ
B Á - - 1.0 ˜
Á ˜
C ÁË - - 0.4˜¯
(ii) Explain whether each of the following is a valid sample path for this
process.
Path 1:
Path 2:
[2]
[Total 4]
A company has for many years offered a car insurance policy with four levels
of No Claims Discount (NCD): 0%, 15%, 30% and 40%. A policyholder who
does not claim in a year moves up one level of discount, or remains at the
highest level. A policyholder who claims one or more times in a year moves
down a level of discount or remains at the lowest level. The company pays
a maximum of three claims in any year on any one policy.
(i) Calculate the premium paid by a policyholder at the 40% discount level
ignoring expenses and profit. [4]
(iii) Calculate the premium paid, in the long term, by a policyholder at the
40% discount level of the policy with protected NCD, ignoring expenses
and profit. [6]
(iv) Discuss THREE issues with the policy with protected NCD which may
each be either a disadvantage or an advantage to the company. [3]
[Total 15]
The solutions presented here are outline solutions for you to use to check
your answers. See ASET for full solutions.
Let N denote the number of audits that are undertaken before the company
changes auditors again. The possible values of N are 2, 3, 4, …. The
associated probabilities are:
P (N = 2) = 1 ¥ 0.2
P (N = 3) = 1 ¥ 0.8 ¥ 0.2
P (N = 4) = 1 ¥ 0.82 ¥ 0.2
and so on.
Hence:
( ) (
E (N ) = 2 (0.2) + 3 (0.8 ¥ 0.2) + 4 0.82 ¥ 0.2 + 5 0.83 ¥ 0.2 + )
(
= 0.2 2 + 3 ¥ 0.8 + 4 ¥ 0.82 + 5 ¥ 0.83 + )
(
E (N ) = 0.2 È 2 + 2 ¥ 0.8 + 2 ¥ 0.82 + 2 ¥ 0.83 +
ÍÎ )
(
+ 0.8 + 0.82 + 0.83 + )
(
+ 0.82 + 0.83 + )
(
+ 0.83 + )
+ ˘˚
This process has a discrete time set (since time is measured by number of
audits), a discrete state space (as described above), and satisfies the
Markov property (since the probability of being in any given state in one time
unit from now depends only on the current state). Hence the process is a
Markov chain.
Keeping the states in the order given above, the transition matrix is:
Ê 0 1 0 0 0 0 ˆ
Á 0 0.8 0.1 0 0.1 0 ˜
Á ˜
Á 0 0 0 1 0 0 ˜
P=Á ˜
Á 0.15 0 0 0.7 0.15 0 ˜
Á 0 0 0 0 0 1˜
Á ˜
Ë 0.05 0 0.05 0 0 0.9¯
(p 1 p 2 p 3 p 4 p 5 p 6 ) P = (p 1 p 2 p 3 p 4 p 5 p 6 )
10
p 3 + 0.7p 4 = p 4 fi p 3 = 0.3p 4 fi p 4 = p3 (4)
3
p1 + p 2 + p 3 + p 4 + p 5 + p 6 = 1 (7)
Ê 10 ˆ
0.15 Á p 3 ˜ + 0.05 (10p 5 ) = p 1
Ë 3 ¯
ie:
ie:
0.5p 3 - 0.5p 1 = p 1 - p 3
fi 1.5p 3 = 1.5p 1
fi p 3 = p1 (10)
10
p4 = p1
3
0.5p 1 + 0.5p 5 = p 1 fi p 5 = p 1
p 6 = 10p 1
Ê 10 ˆ 3
p 1 Á1 + 5 + 1 + + 1 + 10˜ = 1 fi p 1 =
Ë 3 ¯ 64
So the stationary distribution is:
Ê 3 15 3 10 3 30 ˆ
ÁË 64 64 64 64 64 64 ˜¯
Hence the expected proportion of companies using each firm in the long
term is:
3 + 15 18
= = 0.28125 using A
64 64
3 + 10 13
= = 0.203125 using B
64 64
3 + 30 33
= = 0.515625 using C
64 64
This is:
Ê 0.15 0.85 0 0 ˆ
Á 0.15 0 0.85 0 ˜
P= Á ˜
Á 0.03 0.12 0 0.85˜
Á ˜
Ë 0 0.03 0.12 0.85¯
(ii)(a) Probability that a Level 2 policy will be at Level 2 in one year’s time
This is p22 (1) = p22 , ie the same as the one-step transition probability for
staying in state 2, and from the matrix we can see that this is zero.
(ii)(b) Probability that a Level 2 policy will be at Level 2 in two years’ time
This will be the (2, 2) element of the P 2 matrix, which we can obtain by
multiplying the second row of P by the second column of P . This comes
to:
We can construct four equations, one for each of the four long-run
probabilities denoted by p 1, p 2 , p 3 , and p 4 :
We also have:
p1 + p 2 + p 3 + p 4 = 1 (5)
0.15p 4 = 0.85p 3
0.85
fi p4 = p
0.15 3
0.12 ¥ 0.85
p 3 = 0.85p 2 + p3
0.15
0.85
fi p3 = p 2 = 2.65625p 2
0.32
0.85 ¥ 0.85
p4 = p 2 + 0.85p 4
0.32
0.852
fi p4 = p 2 = 15.05208p 2
0.15 ¥ 0.32
Now substitute the above expression for p 3 into equation (1) and we get:
fi p1 =
(0.15 + 0.0796875) p = 0.27022p 2
2
0.85
p1 + p 2 + p 3 + p 4 = 1
fi p 2 = 0.0527
The process can take any integer value, so the state space is the set of
integers: {... - 2, - 1, 0, 1, 2, ...} .
p p p p p p
... -2 -1 0 1 2 ...
(a) Is it aperiodic?
For this process, a return to any state is possible only in an even number of
steps. So each state is periodic with period 2. Hence the process is not
aperiodic.
(b) Is it reducible?
It is possible to reach every state in the process from every other state. So
this process is irreducible.
The state space is infinite. Therefore the process does not have a stationary
distribution.
u +d = m (1)
u -d = j -i (2)
2u = m + j - i
u= 1
2 (m + j - i )
When m is even, the net increase must be even, and when m is odd, the
net increase must be odd. So:
pij( m ) = 0 if m - ( j - i ) is odd
From part (iii), we know that if the process moves from state i to state j
over a period of length m , then the number of upward movements is given
by 21 (m + j - i ) and hence the number of downward movements is:
m- 1
2 (m + j - i ) = 21 (m - j + i )
The probability that any given movement is upward is p . So the probability
of experiencing 1
2 (m + j - i ) upward movements and 1
2 (m - j + i )
downward movements out of a total of m movements is given by the
binomial probability:
Ê m ˆ 1 (m + j - i )
(1 - p ) 2 (
1
m- j +i )
pij( m ) = Á 1 ˜ p2
Ë2( m + j - i )¯
1 p p
0 1 2 ...
p p
1
0 1 2 ...
The probabilities that were zero under the original model will still be zero in
situations (a) and (b) as the states being considered still have period 2, and
the changes haven’t made outcomes that were impossible previously
possible now.
For both situations (a) and (b), if there are not enough time steps available to
include at least one transition out of state 0 within m steps, then the
probabilities are unchanged from before, ie the probabilities pij( m ) are
unchanged for i ≥ m .
pij( m ) will increase for all possible end-states j that include at least one
transition out of state 0 within the sample path. This is because the
probability of transition out of state 0 in the upwards direction is higher than
before. The highest end-state that it is possible to reach that includes a
single visit to state 0 is state m - i .
Given that we are only looking at j > i , we can therefore say that the
non-zero pij( m ) are increased for all values of i < j £ m - i .
pij( m ) will decrease for all possible end-states j , that could have included at
least one transition out of state 0 within the sample path, ie for i < j £ m - i .
This is because the probability of leaving state 0 in the upwards direction
has reduced from before (to zero).
P ( X n = j | X 0 = i 0 , X1 = i1,..., X m -1 = i m -1, X m = i ) = P ( X n = j | X m = i )
for all integer times n > m and all states i0 , i1,..., i m -1, i , j in the state space.
Alternatively, you could describe the Markov property in words. The Markov
property says that the past history of the process is irrelevant. It is only the
current state that affects the transition probabilities.
Alternatively, you could say that the one-step transition probabilities are
independent of time.
1–a 1 2 1–b
P (N = 0) = e -0.3 = 0.74082
P (N = 1) = 0.3e -0.3 = 0.22225
P (N ≥ 2) = 1 - 1.3 e -0.3 = 0.03694
Ê 1 - e -0.3 e -0.3 0 0 ˆ
Á ˜
Á 1 - e -0.3 0 e -0.3 0 ˜
P=Á ˜
-0.3
Á1 - 1.3e 0.3e -0.3 0 e -0.3 ˜
Á ˜
Ë 0 1 - 1.3e -0.3 0.3e -0.3 e -0.3 ¯
Ê 0.25918 0.74082 0 0 ˆ
Á 0.25918 0 0.74082 0 ˜
=Á ˜
Á 0.03694 0.22225 0 0.74082˜
Á ˜
Ë 0 0.03694 0.22225 0.74082¯
pP = p
and:
p 0 + p 25 + p 50 + p 60 = 1
We only need 3 out of these 4 equations, so let’s discard equation (2). From
equation (4), we have:
0.74082p 50 = 0.25918p 60
So:
0.25918
p 50 = p 60 = 0.34986p 60
0.74082
So:
0.12761
p 25 = p 60 = 0.17226p 60
0.74082
So:
0.05757
p0 = p 60 = 0.07771p 60
0.74082
which gives:
p 60 = 0.6251
p 50 = 0.2187
p 25 = 0.1077
p 0 = 0.0486
H0 : N ~ Poi (0.3)
H1 : N ~ Poi (0.3)
(O - E )2
 E
e -0.3 0.3k
130, 200P (N = k | N ~ Poi (0.3)) = 130, 200 ¥
k!
So we have:
(O - E )2
 E
= 0.3265 + 2.8736 + 0.8170 + 4.0544 + 2.5476 = 10.619
If the null hypothesis is true, then the test statistic should come from a c 2
distribution with 5 - 1 - 1 = 3 degrees of freedom. One degree of freedom
has been lost by estimating the Poisson parameter and another has been
lost by constraining the totals to be 130,200.
The examiners also gave full credit to students who didn’t combine the last
two categories. If you did it this way, you would get a test statistic of 25.565,
which you would compare with the c 42 distribution. You would then reject
the null hypothesis at both the 5% and the 1% levels.
(v) Comment
Since the Poisson model is not a good fit, the expected average level of
discount calculated in part (iii) is not likely to be correct. (This is because the
transition probabilities were calculated using the Poisson model.)
The reason that the test statistic is large is mainly due to the fact that we
have more policyholders making multiple claims than expected under the
Poisson model.
These last two observations will tend to push the average discount level in
opposite directions. (More claim-free policies lead to higher average
discount but more multiple claim policies lead to lower average discount.) So
it is difficult to tell what the overall effect would be.
for all times s1 < s2 < < sn < s < t , all states x1, x2, ..., xn and x in S and
all subsets A of S .
(ii) Proof
Markov chains and Markov jump processes are both types of stochastic
process with discrete state spaces and both satisfy the Markov property.
The difference between them is that a Markov chain has a discrete time set
and a Markov jump process has a continuous time set.
(iii)(b) Irreducibility
Let X (t ) denote the floor that the lift is on at time t . The state space
consists of the 5 floors in the office, which we can label {0,1, 2, 3, 4} .
Since it is possible to get from any floor to any other floor using the lift, an
irreducible model is appropriate.
We could get round this problem by using a more complicated state space,
eg defining the state space to represent the current floor and a number of
past floors.
and:
p A + p B + pC = 1 (1)
and:
From (4):
0.25p A = 0.5p C
fi p A = 2p C
2p C + 3p C + p C = 1
1
So p C = and the stationary distribution is:
6
Ê1 1 1ˆ
ÁË 3 2 6 ˜¯
(iii) Comment
The stationary distribution gives the long-term probabilities that a car will be
returned to each of the three locations: airport, beach and city. However, it
does not tell us how likely a car is to be picked up from each of the three
locations. This is more important than the stationary distribution for drop-off
points.
Other factors that should be considered when deciding how many cars
should be based at each location include the available space and facilities at
each location, the number of cars owned by the company and the cost of
transporting a car from one location to another.
(iv) Sketch
1
0.9
0.8
0.7
Probability
Airport
0.6
0.5 Beach
0.4 City
0.3
0.2
0.1
0
0 1 2 3 4 5 6 7 8 9 10
Number of times car has been rented
The graph shows that the long-run probabilities for the airport, the beach and
1 1 1
the city are , and , respectively. Note that the beach and city
3 2 6
probabilities are the same at times 0 and 1.
B C
The third bullet point tells us that the transitions A→B and A→C are both
possible, each with the same probability, p say.
p p
B C
The question doesn’t actually say that the probability p cannot equal zero,
but we can easily see that it can’t equal zero. If it did, it would not be
possible to move from A to either of the other two states, which would make
A an absorbing state and this would contradict the irreducible property.
Since the process is periodic and it only has 3 states, the period must be
either 2 or 3. If the period were 3, the process would have to consist of a
triangular ‘one-way system’ going in one direction or the other:
A A
B C B C
However, these models are not consistent with the two ‘p’ arrows we
established earlier, which both go out of A. So the process must have a
period of 2.
Since it is periodic, we also know that none of the states have loops going
back to themselves. So the process must have the following form:
p p
B C
Since there is only one exit route from each of states B and C, the
associated probabilities must equal 1. Since there are two exit routes from
state A with equal probability (and staying in A is not possible), we also know
that p = ½ .
½ ½
1 1
B C
p1 + p2 +
G
p2 + p0
p1
S
p2 + p0
p1
B
p2 + p0
p0 + p1
O
The process has discrete states and operates in discrete time, with
movements occurring only at the end of each calendar year.
This set-up has the Markov property because all the members in a particular
state in a particular year have the same probabilities for moving between the
states in the future.
p = pP , Âpi = 1, pi ≥ 0
i
This chain has a finite state space, with 4 states. So there is at least one
stationary distribution.
The chain is also irreducible, ie it is possible to move from any state i to any
state j (eventually). We can see this because the chain allows a complete
circuit through the states O→B→S→G→S→B→O). So, in fact, there is a
unique stationary distribution.
O B S G
O È p0 + p1 p2 + 0 0 ˘
Í ˙
B Í p0 p1 p2 + 0 ˙
S Í 0 p 0 p1 p2 + ˙
Í ˙
G ÍÎ 0 0 p0 p1 + p2 + ˙˚
We are told that p0 = 0.4 , p1 = 0.4 and p2 + = 0.2 . So the transition matrix
for the model is:
O B S G
O È0.8 0.2 0 0 ˘
Í ˙
B Í 0.4 0.4 0.2 0 ˙
S Í 0 0.4 0.4 0.2 ˙
Í ˙
G ÍÎ 0 0 0.4 0.6 ˙˚
The stationary distribution is the set of probabilities p i that satisfy the matrix
equation p = p P with the additional condition Âpi = 1.
i
Written out in full, the matrix equation for the stationary distribution
(p O , p B , p S , p G ) satisfies the system of equations:
So:
So we have:
1 8
1.875p O = 1 fi p O = =
1.875 15
( 8 4 2 1
, , ,
15 15 15 15 )
(iv) Profitability of scheme
4 + 20 ¥ 2 + 30 ¥ 1 = 7 1
0p O + 10p B + 20p S + 30p G = 0 + 10 ¥ 15 15 15 3
0 ¥ pO + 10 ¥ p1 + 10n ¥ p2 +
¤ 4 + 2n > 7 31
¤ n > 1 32
But, since n must be at least 2, this will always be true. So the expected
net profit of the scheme in the long-term will be positive.
Ê 0.4 0 0 0.6ˆ
Á 0.4 0.4 0 0.2˜
Á ˜
Á 0.2 0.4 0.4 0 ˜
Á ˜
Ë 0 0.2 0.4 0.4¯
Ê 0.4 0 0 0.6ˆ
Á 0.4 0.4 0 0.2˜
(p 1 p2 p3 p4)Á ˜ = (p 1 p 2 p 3
Á 0.2 0.4 0.4 0 ˜
p4)
Á ˜
Ë 0 0.2 0.4 0.4¯
3
(3) gives 0.4p 4 = 0.6p 3 fi p 4 = p3
2
7
0.4p 2 + 0.4p 3 + 0.3p 3 = p 2 fi 0.6p 2 = 0.7p 3 fi p 2 = p3
6
Ê7 ˆ 2 10
0.4p 1 + 0.4 Á p 3 ˜ + 0.2p 3 = p 1 fi 0.6p 1 = p 3 fi p 1 = p
Ë6 ¯ 3 9 3
Ê 10 7 3ˆ
p3 Á + + 1+ ˜ = 1
Ë 9 6 2¯
which gives:
9 10 21 27
p3 = , p1 = , p2 = , p4 =
43 43 86 86
Ê 10 21 9 27 ˆ
So the stationary distribution is Á .
Ë 43 86 43 86 ˜¯
If the shop has 1 snake in stock at the start of a day, the probability it
will have to restock before it opens again the next day is 0.6.
If the shop has 2 snakes in stock at the start of a day, the probability it
will have to restock before it opens again the next day is 0.2.
If the shop has 3 or 4 snakes in stock at the start of a day, it will not
have to restock before it opens again the next day.
10 21 81
0.6p 1 + 0.2p 2 = 0.6 ¥ + 0.2 ¥ = = 0.1884
43 86 430
If the shop has 1 snake in stock at the start of a day, it loses 1 potential
sale with probability 0.2.
If the shop has 2 or more snakes in stock at the start of a day, there are
no lost sales that day.
So the expected long-term number of sales lost per trading day is:
10 2
0.2p 1 = 0.2 ¥ = = 0.0465
43 43
(v) Expected restocking orders and lost sales under new set up
The shop will restock if the number of snakes at the end of the day is 0 or 1.
So the number of snakes in stock at the start of each trading day is now 2, 3
or 4. The new stationary distribution is found from:
Ê 0.4 0 0.6ˆ
( p2 p3 p4 ) Á 0.4 0.4 0.2˜ = ( p2
Á ˜
p3 p4 )
ÁË 0.2 0.4 0.4˜¯
and p2 + p3 + p4 = 1
3
From (6), 0.4 p4 = 0.6 p3 fi p4 = p3
2
7
0.4 p2 + 0.4 p3 + 0.3 p3 = p2 fi 0.6 p2 = 0.7 p3 fi p2 = p3
6
Ê7 3ˆ
Now, using the fact that p2 + p3 + p4 = 1 , we have p3 Á + 1 + ˜ = 1
Ë6 2¯
So, the stationary distribution under the new restocking arrangement is:
Ê 7 3 9 ˆ
ÁË 22 11 22 ˜¯
The expected long-term number of restocking orders per trading day is now:
7 3 27
0.6 p2 + 0.2 p3 = 0.6 ¥ + 0.2 ¥ = = 0.2455
22 11 110
No sales will be lost because the probability of selling more than 2 snakes in
a day is 0.
Restocking when there are 2 or more snakes in stock means that there will
be no lost sales. This can also be achieved by restocking when the number
of snakes falls to 1 as we have seen in (v) (because the probability of selling
more than 2 snakes in a day is 0).
If the shop is restocked only when there are no snakes in stock, the
expected long-term cost per day is:
81 2
C+ P
430 43
If the shop is restocked when the number of snakes falls below 2, the
expected long-term cost per day is:
27
C
110
So the owner should change to restocking when the number of snakes falls
below 2 if:
27 81 2
C< C+ P
110 430 43
22
ie C< P
27
If a Markov chain has a finite state space, ie if the number of states is finite,
then it has at least one stationary distribution.
If, in addition to having a finite state space and being irreducible, a Markov
chain is also aperiodic, ie if there does not exist an integer d > 1 such that
return to a given state is only possible in multiples of d steps, then it will
converge to the unique stationary distribution.
If the current state is 1 ‘Poor’, then, according to the rules, there is:
a 20% chance the process will move up one level to 2 ‘Satisfactory’
a 20% chance the process will move down one level (which will leave it
on 1 ‘Poor’, since it is already on the lowest level)
a 10% chance the process will move up two levels to 3 ‘Good’
a 10% chance the process will move down two levels (which will leave it
on 1 ‘Poor’, since it is already on the lowest level).
Analysing each of the other states in a similar way leads to the following
one-step probability transition matrix:
1 2 3 4
1 È0.7 0.2 0.1 0 ˘
Í ˙
P= 2 Í0.3 0.4 0.2 0.1˙
3 Í 0.1 0.2 0.4 0.3 ˙
Í ˙
4 ÎÍ 0 0.1 0.2 0.7 ˚˙
We know that each child starts on level 2 ‘Satisfactory’ on the first Friday.
The probabilities for the third Friday are the probabilities after two steps.
These can be found from the second row of the matrix P 2 .
È ∑ ∑ ∑ ∑ ˘
Í ˙
0.35 0.27 0.21 0.17
=Í ˙
Í ∑ ∑ ∑ ∑ ˙
Í ˙
ÎÍ ∑ ∑ ∑ ∑ ˚˙
qx
mˆ x + f =
E xc
From the table we can see that lives 3, 6 and 10 died before their 1st
birthdays. So, for the year of age (0,1) , the number of deaths is:
q0 = 3
The central exposed to risk for the year of age (0,1) for the lives who are
recorded as deaths is:
2 months for life 3
9 months for life 6
1 month for life 10.
We are told that the remaining 7 lives (with no date of death shown) survived
until their 1st birthday. So their exposed to risk is a full year.
So the total central exposed to risk for the year of age (0,1) is:
Since we are assuming a constant force of mortality over the year, the
estimate of the force of mortality for the year of age (0,1) applies to age 0.5
and is equal to:
q0 3
mˆ 0.5 = = = 0.375
E0c 8
Since we are assuming that the force of mortality is constant over the year,
the values of q and m are related by the equation:
q 0 = 1 - p 0 = 1 - exp Ê - Ú ms ds ˆ = 1 - e - m0.5
1
Ë 0 ¯
With this model, individuals in State L who reduce their order have different
probabilities for their future movements, depending on whether they:
(a) increased/maintained or
(b) reduced their order the previous year. They will drop by either one or
two levels, depending on their previous actions.
So this violates the Markov property, which requires that the current state
fully determines the future probabilities.
To preserve the Markov property, we need to split State L into two separate
states, L+ and L– (say), where:
L+ = Members paying the Loyalty price who increased or maintained
their order the previous year
L– = Members paying the Loyalty price who reduced their order the
previous year.
L+ L-
B G L+ L- S
B È0.4 0.6 0 0 0 ˘
Í ˙
G Í 0.3 0.1 0.6 0 0 ˙
L+ Í 0 0.3 0.1 0 0.6 ˙
Í ˙
L - Í0.3 0 0.1 0 0.6 ˙
S ÍÎ 0 0 0 0.3 0.7 ˙˚
The stationary distribution is the set of probabilities p i that satisfy the matrix
equation p = p P with the additional condition Âpi = 1.
i
p L - = 0.3p S
fi 0.12p S = 0.6p L +
0.12
fi p L+ = p S = 0.2p S
0.6
fi 0.15p S = 0.6p G
0.15
fi pG = p S = 0.25p S
0.6
fi 0.6p B = 0.165p S
0.165
fi pB = p S = 0.275p S
0.6
fi 2.025p S = 1
1 40
fi pS = =
2.025 81
(p B , p G , p L + , p L - , p S ) = ( 11 10 8 12 40
, , , ,
81 81 81 81 81 )
(iii)(c) Long-run average price
We can calculate the long-run average price paid by applying the long-run
probabilities to the price paid in each state:
= 81 13
27
% or 81.48%
Let X t denote the state the process is in at time t . The required probability
is:
pL,L (2) = P ( X t + 2 = L | X t = L )
pL,L (2) = P ( X t + 2 = L | X t = L + or X t = L - )
pL,L (2) = P ( X t + 2 = L | X t = L + ) ¥ P ( X t = L + )
+P ( X t + 2 = L | X t = L - ) ¥ P ( X t = L - )
To evaluate this probability, we will assume that the customer base has
settled down with the long-run probabilities we worked out in part (iii)(b).
8
This means that the proportions in States L+ and L– will be p L + = 81
and
p L- = 12 , so that a customer in one of these two ‘L’ states has a probability
81
of 0.4 of being in State L+ and 0.6 of being in State L–.
p L+
pL,L (2) = P ( X t + 2 = L | X t = L + ) ¥
p L+ + p L-
p L-
+P ( X t + 2 = L | X t = L - ) ¥
p L+ + p L-
= P ( X t + 2 = L | X t = L + ) ¥ 0.4 + P ( X t + 2 = L | X t = L - ) ¥ 0.6
To work out the two-step probabilities in this expression, we can follow the
possible paths in the transition diagram:
and P ( X t + 2 = L | X t = L - ) = P ( X t + 2 = L + | X t = L - ) + P ( X t + 2 = L - | X t = L - )
= {P (L -, L +, L + )} + {P (L -, S, L - )}
= 0.12 + 0.6 ¥ 0.3
= 0.19
The required probability is then pL,L (2) = 0.37 ¥ 0.4 + 0.19 ¥ 0.6 = 0.262 .
There were several other acceptable approaches you could have used here:
(1) You can assume that Loyalty customers are equally likely to start in
either State L+ or L–, which gives a probability of
0.37 ¥ 0.5 + 0.19 ¥ 0.5 = 0.28 .
(2) You can express the answer as 0.37 p + 0.19(1 - p ) = 0.19 + 0.18 p ,
where p is the unknown proportion originally in state L+.
(3) You can say that the probability will lie somewhere between 19% and
37%, depending on the proportions originally in the L+ / L– states.
Note that adding together the four probabilities 0.19, 0.18, 0.01 and 0.18
does not give the right answer here. This is calculating the probability that
an L+ customer is paying the Loyalty price in 2 years’ time PLUS the
probability that an L– customer is paying the Loyalty price in 2 years’ time.
Changes in the pattern of hay consumption might change the total amount of
hay required by customers.
A competitor might introduce a rival scheme that takes away some types of
customers.
The relative price of hay from different sources (eg cheap imports) might
affect his total sales.
State 0
State 0 means that the actuary has no umbrellas at his current location. So
he is certain to leave without one and there will be 2 at his destination. This
is represented by the arrow from state 0 to state 2 labelled with probability 1.
State 2
State 2 means that the actuary has 2 umbrellas at his current location. If it is
raining (which has probability p ) he will take 1 with him, so that there will
then be 1 at his destination. If it is not raining (which has probability 1 - p )
he will not take an umbrella and there will be none at his destination. This is
represented by the two arrows leading out of state 2.
State 1
State 1 means that the actuary has 1 umbrella at his current location (which
means that there is already 1 at his destination). If it is raining (which has
probability p ) he will take 1 umbrella with him, so that there will be 2 at his
destination. If it is not raining (which has probability 1 - p ) he will not take
an umbrella and there will be 1 at his destination. This is represented by the
two arrows leading out of state 1.
0 1 2
0 È 0 0 1˘
Í ˙
1 Í 0 1 - p p ˙
2 ÍÎ1 - p p 0 ˙˚
È 0 0 1˘ È 0 0 1˘
Í
2 ˙ Í ˙
P =Í 0 1- p p˙ Í 0 1- p p˙
ÍÎ1 - p p 0 ˙˚ ÍÎ1 - p p 0 ˙˚
È 1- p p 0 ˘
Í 2 2 ˙
= Í p(1 - p ) (1 - p ) + p p(1 - p ) ˙
Í ˙
ÍÎ 0 p(1 - p ) 1 - p + p 2 ˙˚
The stationary distribution is the set of probabilities p i that satisfy the matrix
equation p = p P with the additional condition  pi = 1.
i
Written out in full, this matrix equation is:
p 0 = (1 - p )p 2 (1)
p 1 = (1 - p )p 1 + pp 2 (2)
p 2 = p 0 + pp 1 (3)
(p 0 , p 1, p 2 ) = ((1 - p )p 2, p 2, p 2 )
(1 - p )p 2 + p 2 + p 2 = 1
fi (3 - p )p 2 = 1
1
fi p2 =
3-p
Ê 1- p 1 1 ˆ
The stationary distribution is therefore (p 0 , p 1, p 2 ) = Á , , .
Ë 3 - p 3 - p 3 - p ˜¯
1- p p(1 - p )
p0 ¥ p = ¥p=
3-p 3-p
In this model there are two states for the weather, say R = ‘raining’ and N =
‘not raining’.
R N
R Èr 1- r ˘
Í ˙
N Îs 1 - s ˚
However, in this new situation, state 2, for example, will include journeys
when it rained on the previous journey and ones when it didn’t. The
probabilities for moving to state 1 next in these two cases will now be r
and s , which are different. So this violates the Markov property.
To preserve the Markov property, we need to incorporate the weather for the
previous journey in the states. This would give us 6 states, eg 2R = ‘2
umbrellas and it was raining on the previous journey’. However, there will
not be a 0R state because the actuary can only enter state 0 from state 2,
and if it was raining on the previous journey he would have brought an
umbrella with him.
s
r
2R 1R 1N 1–s
1–r
r
s 1–r s
1–s
0N 2N
1–s
n +1 n
X n +1 = Â Yj = Â Yj + Yn +1 = X n + Yn +1
j =1 j =1
We also know that the probabilities for Yn +1 are determined based only on
1Ê X ˆ
X n using the formula P(Yn +1 = 1) = 1 + n ˜ . So the value of X n fully
4 ÁË n ¯
determines the probability distribution for X n +1 , and hence also for
X n + 2 , X n +3 , , ie the series X i has the Markov property.
The probabilities for the value of Yn +1 , on the other hand, depend on the
value of X n , which cannot be determined based only on the value of Yn
(apart from in year 1 when the probability is always 1 ). So the series Yi
2
does not have the Markov property.
(ii)(a) Is X i irreducible?
(ii)(b) Is X i time-homogeneous?
X0 X1 X2 X3
1 3 3
2 4 4
0 0 0 0
1 1
4 4
1
2 1 1
1 1
5
2 8 3
8
1
2 2 2
1
2
1
2 3
1
P( X1 = 1) = P(Y1 = 1) =
2
1Ê X ˆ
and P( X n +1 - X n = 1) = P(Yn +1 = 1) = 1 + n ˜ , n = 1,2,3,
4 ÁË n ¯
(iv) Probability
The probability that the driver has accidents during exactly two of the first
three time periods is P ( X 3 = 2) . This can be calculated by adding the
probabilities for the various paths going from X 0 = 0 to X 3 = 2 :
P ( X 3 = 2) = P (0 Æ 0 Æ 1 Æ 2) + P (0 Æ 1 Æ 1 Æ 2) + P (0 Æ 1 Æ 2 Æ 2)
= ( 21 ¥ 41 ¥ 83 ) + ( 21 ¥ 21 ¥ 83 ) + ( 21 ¥ 21 ¥ 21 )
= 3 + 3 + 1
64 32 8
= 17 ( = 0.2656)
64
(v) Comment
With this model, the effect of a past accident reduces over time (because it is
averaged over all past years), which also seems reasonable.
This model doesn’t take into account – not directly, at least – many other
factors that are likely to have a big effect on the rate of future accidents,
such as the driver’s age and experience, their annual mileage and where
they live and work.
This model only considers ‘no accidents’ versus ‘at least one accident’. The
company might want to take into account the actual number of accidents
each year, since accident prone drivers may have several accidents in the
same year.
We also need to distinguish between members who are in their first and
second months of sickness because the number of months for which they
can continue receiving the full rate of sickness pay is different for the two
groups (which means the model wouldn’t be Markov if we combined them).
0.9
0.25
0.1 0.25 0.25
H S1 S2 S3+
0.75
0.75
0.75
H S1 S2 S3 +
H È 0.9 0.1 0 0 ˘
Í ˙
S1 Í 0.75 0 0.25 0 ˙
S2 Í0.75 0 0 0.25˙
Í ˙
S3 + ÍÎ0.75 0 0 0.25˙˚
The stationary distribution is the set of probabilities p i that satisfy the matrix
equation p = p P with the additional condition Âpi = 1.
i
(
p S1 = 0.1p H = 1 p
10 H )
Equation (3) then gives:
0.75p S 3+ = 0.25p S 2
fi p S3+ =
0.25
p =
0.25
0.75 S 2 0.75
(0.025p H ) = 0.00833p H = ( 1 p
120 H )
So the vector of stationary probabilities has the form:
( 1 p , 1 p , 1 p
(p H , p S1, p S 2 , p S 3+ ) = p H , 10 H 40 H 120 H )
p H + 10
1 p +
H
1 p
40 H
1 p =1
+ 120 H
fi 17 p =1
15 H
fi pH = 15
17
fi xp H ≥ 0p H + 1p S1 + 1p S 2 + 0.5p S 3+
15 x≥ 3 3 +
+ 136 1 1 =
¥ 136 31
17 34 2 272
fi x≥ 17 ¥ 31 = 31 = 0.1292
15 272 240
y p H ≥ 0p H + 1p S1 + 1p S 2 + 1p S 3+
15 y≥ 3 3 + 1 =
+ 136 2 = 0.1176
17 34 136 17
fi y≥ 17 2 =
¥ 17 2 = 0.1333
15 15
(vii) Comment
The reduction to 50% from the third month results in a slightly lower
contribution rate, ie 12.92% compared to 13.33%. In terms of the actual
amounts paid, this is a reduction of only 3%. This is a small change
because the probability of being sick for 3 months in a row is relatively small
in this model.
Employees may have more peace of mind knowing that they will continue to
receive their full pay. However, this may also reduce their incentive to return
to work if they are well enough, since they will receive the same amount of
pay anyway.
Let p be the probability of making exactly one claim in the year. Then:
Ê 1.25 p 1 - 1.25 p 0 ˆ
Á
P = 1.25 p 0 1 - 1.25 p˜
Á ˜
ÁË 0.25 p p 1 - 1.25 p˜¯
This Markov chain has a finite state space, is irreducible and is aperiodic.
This means that the long-term behaviour of the chain can be obtained by
finding the unique solution p to the equation:
pP = p
Ê 1.25 p 1 - 1.25 p 0 ˆ
¤ p Á 1.25 p 0 1 - 1.25 p˜ = p (1)
Á ˜
ÁË 0.25 p p 1 - 1.25 p˜¯
where p = (1 - 2a, a, a ) .
So, we have:
Ê 5p ˆ
2 Á1 - =1
Ë 4 ˜¯
1 4 2
¤ p= ¥ =
2 5 5
Ê 0.5 0.5 0 ˆ
P = Á 0.5 0 0.5˜
Á ˜
ËÁ 0.1 0.4 0.5˜¯
(1 - 2a )(1 - 1.25 p ) + ap = a
1 2a
¤ -a+ =a
2 5
1 5 5
¤ a= ¥ =
2 8 16
● pj = Â p i pij
i ŒS
● pj ≥0
● Â pj =1
j ŒS
Ê 0.4 0.6 ˆ
P=Á
Ë 0.05 0.95¯˜
This Markov chain has a finite state space, is irreducible and is aperiodic.
This means that the long-term behaviour of the chain can be obtained by
finding the unique solution p to the equation:
pP = p
Ê 0.4 0.6 ˆ
¤ (p 1, p 2 ) ÁË 0.05 0.95˜¯
= (p 1, p 2 )
0.4p 1 + 0.05 (1 - p 1 ) = p 1
¤ 0.05 = 0.65p 1
1 12
¤ p1 = and, hence p 2 = .
13 13
Ê 0.2 0.8 ˆ
P=Á
Ë 0.05 0.95¯˜
Solving p P = p , we get:
Ê 0.2 0.8 ˆ
(p 1, p 2 ) ÁË 0.05 0.95˜¯
= (p 1, p 2 )
0.2p 1 + 0.05 (1 - p 1) = p 1
¤ 0.05 = 0.85p 1
1 16
¤ p1 = and, hence p 2 = .
17 17
Solving for the maximum that the stadium should be prepared to pay $M ,
we get:
1 1
¥ $11,000 > ¥ ($10,000 + $M )
13 17
¤ $M < $4,384.62
So, the maximum that the stadium should be prepared to pay is $4,384.62.
The boundaries of this process are the 0% state and the 40% state. Both
are mixed because, for example, a policyholder in the 40% state may either
remain in the 40% state for the next year or move to a different state (25% or
10%) following a claim.
To model this as a Markov chain, we require 4 states (0%, 10%, 25% and
40%).
p0 p0 p0 p0
0% 10% 25% 40%
p1 + p2 + p1 + p2 + p1 p1
p2 +
where:
The number of claims made during the year will have a Bin(12, 0.04)
distribution. So:
p0 = 0.9612 = 0.61271
Ê12ˆ
p1 = Á ˜ ¥ 0.9611 ¥ 0.04 = 0.30635
Ë 1¯
and:
p2 + = 1 - p0 - p1 = 0.08094
The long-term probabilities for the 4 states can be found by finding the
stationary distribution, ie the set of probabilities p i that satisfy the matrix
equation p = p P with the additional condition  pi = 1.
i
0.61271p 25
p 40 = = 1.5820p 25
1 - 0.61271
p 10 - 0.30635p 25 - 0.08094p 40
p0 =
0.61271
0.84111p 25 - 0.30635p 25 - 0.08094 ¥ 1.5820p 25
=
0.61271
= 0.66379p 25
fi 4.0869p 25 = 1
fi p = 1
25 = 0.24468
4.0869
So the proportion of policyholders at the 25% level in the long run is 24.5%.
U C D Total
U I II III 6
C II III II 7
D II III I 6
U C D U C D
1 2 3 1 1 1
U 6 6 6 U 6 3 2
or (if we cancel the fractions)
C 2 3 2 C 2 3 2
7 7 7 7 7 7
D 2 3 1 D 1 1 1
6 6 6 3 2 6
We can find this probability by listing the outcomes for the next 3 matches
where there are at least 2 wins for United, noting that the last match that was
played was a ‘C’. In the table below ‘*’ indicates a ‘wild card’, ie any result.
(C)UU* 21 2
7 6 76
(C)UCU 2 62 72 7867
7
(C)UDU 2 63 62 712
7 66
(C)CUU 3 72 61 7766
7
(C)DUU 2 62 61 7646
7
2 14 16 1 1 4 10
0.15873
7 6 7 7 6 7 6 6 21 21 63 63
As this process is irreducible, all the states have the same period. We can
see that the process is aperiodic because it is possible to ‘loop’ from the 0%
state back to itself in one time step, so this state, and hence all the states,
has a period of 1.
The stationary distribution is the set of probabilities i that satisfy the matrix
equation P with the additional condition i 1.
i
50 0.8 25
60 0.8(0.8 25 ) 0.8 60
0.2 60 0.64 25
60 3.2 25
25 0.8 0 0.2(3.2 25 )
0.36
0 25 0.45 25
0.8
5.45 25 1
1
25 0.1835
5.45
The 0% and 60% states are aperiodic because they have ‘loops’. The 25%
and the 50% states are also aperiodic because they interconnect with the
0% state and therefore have the same periodicity. So the whole chain is
aperiodic.
In the long run, all policyholders will now end up in the 60% state. So the
stationary distribution for the protected scheme is
( 0 , 25 , 50 , 60 ) 0,0,0,1 .
e - mE ( m E )di
P (Di = di ) =
di !
Assuming that there are i independent lives in the population, the likelihood
function for the whole population is therefore:
n n
n e - mE ( m E )di  di  di n 1
L=’ = e - n m E ¥ m i =1 ¥ E i =1 ¥ ’
i =1 di ! i =1 di !
n
ln L = - n m E + Â d i ln m + ... (terms not involving m )
i =1
∂ 1 n
ln L = - nE + Â di
∂m m i =1
1 n
m =
nE
 di
i =1
∂2 1 n
ln L = - Â di <0
∂m 2 m2 i =1
We can calculate the waiting times for each student, ie the time between
arriving at the bus stop and the next bus arriving or the student leaving.
We can now calculate the MLE of the arrival rate for the buses by dividing
the total number of buses arriving by the total waiting time for the students.
(Note that only 4, not 6, buses arrived.) This gives:
4 4
mˆ = per minute = = 1.33 per hour
180 3
Four buses arrived during the period 4pm to 6.10pm. However, there are
gaps in this period during which no-one was at the bus stop, so the length of
the observation period is not the full 2 hours and 10 minutes. For example, if
a bus had arrived at 4.07pm, its arrival would not have been recorded.
There is at least one person waiting at the bus stop during the following
intervals:
115
So the bus stop was ‘observed’ for a total of 115 minutes or hours
60
between 4pm and 6.10pm. So the estimated arrival rate of buses is:
4 48
= = 2.087 per hour
115 / 60 23
Note that dividing the number of students who caught a bus by the total
observed waiting time for all the students gives the answer:
6 6
= = 2 per hour
180 / 60 3
This is an estimate of the rate at which students catch a bus, rather than an
estimate of the arrival rate of buses.
(iii) Comment
The assumption that the arrival of buses follows a Poisson process may not
be valid. In particular:
The model assumes that the arrival rate is constant. In reality there
may be more buses scheduled to arrive during the busiest times of day.
The model also assumes that the arrival times of the buses are
random. In reality there is a timetable, so they will be most likely to
arrive around those times.
The model also assumes that the times between buses are
independent. In reality, several consecutive buses may all be affected
by an obstruction such as a traffic accident and the times would not be
independent.
Also, there were periods when no students were at the bus stop, eg between
4:50 and 4:55. We do not know whether any buses arrived during these
periods. These would affect our estimate of the arrival rate.
(iii) Probability
For brevity we will refer to the states by their initial letters (eg E = Excellent).
We can derive the probability of moving from state E to state P in two steps
by considering the possible paths:
0.5
¥ 0.2
+ 0.2
¥ 0.3
+ 0.2
¥ 0.5
= 0.26
E ÆE ÆP E ÆGÆP E ÆP Æ P
The long-term probabilities for the 3 states can be found by finding the
stationary distribution, ie the set of probabilities p i that satisfy the matrix
equation p = p P with the additional condition Âpi = 1.
i
Written out in full, this matrix equation is:
0.5p G = 0.2p E
fi p G = 0.4p E
(p E , p G , p P ) = (p E ,0.4p E ,0.64p E )
p E + 0.4p E + 0.64p E = 1
fi 2.04p E = 1
fi pE =
1
2.04
= 0.4902 = ( 25
51 )
So the long-term proportions are:
The kiln is fired 100 times a year and the cost will be £1,000 + £50 = £1,050
whenever it finishes the cycle in the Failed state. So the expected annual
cost will be:
To deal with this new replacement policy, we need to combine both the
Failed and the Poor columns in the original matrix with the Excellent column,
which gives the following 2 ¥ 2 matrix:
È0.8 0.2˘
P* = Í ˙
Î0.5 0.5 ˚
p E* = 0.8p E* + 0.5p G*
p G* = 0.2p E* + 0.5p G*
This gives:
p E* = 5
7
= 0.7143 and p G* = 2
7
= 0.2857
0.15
0.85 1N 2R
0.3
0.7 0.75
0.25
1P 2N
0.85
0.15
1N 1P 2R 2N
1N È0.85 0 0.15 0 ˘
Í ˙
1P Í 0.7 0 0.3 0 ˙
2R Í 0 0.25 0 0.75 ˙
Í ˙
2N ÎÍ 0 0.15 0 0.85 ˚˙
As this process is irreducible, all the states have the same period. We can
see that the process is aperiodic because it is possible to ‘loop’ from the 1N
state back to itself in one time step, so this state, and hence all the states,
have a period of 1.
The table below shows the probabilities of moving from State 1P to State 2R
in different numbers of seasons.
Cumulative
Seasons Path Probability
probability
1 season 1P Æ 2R 0.3 0.3
2 seasons 1P Æ 1N Æ 2R 0.7 ¥ 0.15 = 0.105 0.405
3 seasons 1P Æ 1N Æ 1N Æ 2R 0.7 ¥ 0.85 ¥ 0.15 = 0.08925 0.49425
3 2
4 seasons 1P (Æ 1N ) Æ 2R 0.7 ¥ 0.85 ¥ 0.15 = 0.07586 0.57011
5 seasons 1P (Æ 1N )4 Æ 2R 0.7 ¥ 0.853 ¥ 0.15 = 0.06448 0.63460
So we see that it will take 5 seasons for the probability to exceed 60%.
A chain is Markov if the current state fully determines the probabilities for
future movements between states.
With this NCD system, if policyholders on Level 2 make a claim, the state
they move to will depend on whether or not they made a claim the previous
year. If they made no claim the previous year, they move down to Level 1,
but if they made at least one claim in the previous year, they move down to
Level 0. So, to obtain a Markov model, we need to split Level 2 into two
states, eg 2+ for those who didn’t make a claim the previous year and 2– for
those who did.
p p p 1–p
p
p 2-
0 1 2+ 2- 3
0 Èp 1- p 0 0 0˘
Í ˙
1 Íp 0 1- p 0 0 ˙
2+ Í0 p 0 0 1- p ˙
Í ˙
2- Í p 0 0 0 1- p ˙
3 Í0 0 0 p 1 - p ˙˚
Î
The stationary distribution is the set of probabilities p i that satisfy the matrix
equation p = p P with the additional condition Âpi = 1.
i
p0 = pp 0 + pp 1 + pp 2 - ... (1)
p1 = (1 - p )p 0 + pp 2 + ... (2)
p 2+ = (1 - p )p 1 ... (3)
p 2- = pp 3 ... (4)
p3 = (1 - p )p 2 + + (1 - p )p 2 - + (1 - p )p 3 ... (5)
p 2 - = pp 3
p 3 = (1 - p )p 2 + + (1 - p )( pp 3 ) + (1 - p )p 3
p 2 + = (1 - p )p 1
p2
fi p 3 = (1 - p)p 1
1- p
p2
fi p1 = p3 ... (7)
(1 - p )2
p2 1
2
=
(1 - p ) 9
p 1
=
1- p 3
fi 3p = 1 - p
fi 4p = 1
p= 1
fi 4
Substituting this value of p into Equations (4), (6) and (7), we get:
p 1 = 91 p 3
p 2 - = 41 p 3
p 2+ = 1 p
12 3
p 0 = pp 0 + pp 1 + pp 2 -
ie p 0 = 41 p 0 + 41 p 1 + 41 p 2 -
fi 4p 0 = p 0 + p 1 + p 2 -
fi 3p 0 = p 1 + p 2 -
fi p 0 = 31 (p 1 + p 2 - ) = (
1 1
p
3 9 3 )
+ 41 p 3 = 13
p
108 3
(p 0 , p 1, p 2 + , p 2 - , p 3 ) = ( 13
p , 1 p , 1 p , 1 p ,p
108 3 9 3 12 3 4 3 3 )
13
p
108 3
+ 91 p 3 + 12
1 p + 1p +p = 1
3 4 3 3
169
fi p
108 3
=1
108
fi p3 = 169
= 0.6391
(p 0 , p 1, p 2 + , p 2 - , p 3 ) = ( 13 12
, , 9 , 27 , 108
169 169 169 169 169 )
or
(0.0769,0.0710,0.0533,0.1598,0.6391)
The proportion at the 25% discount level in the long run is:
9 27 36
p 2+ + p 2- = 169
+ 169 = 169
= 0.2130
So we will need to split Level 1 into two states, eg State 1+ for those who
didn’t make a claim the previous year and State 1– for those who did. So we
now need 6 states in total.
Let q denote the probability of a claim if no claim was made the previous
year and r denote the probability of a claim if a claim was made the
previous year.
0 1+ 1- 2 + 2 - 3
0 È r 1- r 0 0 0 0 ˘
Í ˙
1+ Íq 0 0 1- q 0 0 ˙
1- Í r 0 0 1- r 0 0 ˙
Í ˙
2 + Í0 0 q 0 0 1- q˙
2 - Ír 0 0 0 0 1- r ˙
Í ˙
3 ÎÍ 0 0 0 0 q 1 - q ˚˙
Since users are equally likely to move to any of the possible states, the
probabilities are as shown below:
1
2
N(ile) B(anana)
1
2
1
1 1 2
2 3
1
3
1
C(heep) H(andbook)
1
3
N B C H
N È0 1 1 0˘
Í 2 2 ˙
B Í1 0 1
0˙
Í2 2 ˙
C Í1 1 0 1˙
Í3 3 3˙
H ÍÎ 0 0 1 0 ˙˚
The long-term probabilities for the four states can be found by calculating the
stationary distribution, ie the set of probabilities p i that satisfy the matrix
equation p = p P , with the additional condition  pi = 1.
i
Written out in full, this matrix equation is:
pN = 1p + 31 p C
2 B
pB = 1p + 31 p C
2 N
pC = 1p + 21 p B + p H
2 N
1
pH = p
3 C
Equivalently:
6p N = 3p B + 2p C (1)
6p B = 3p N + 2p C (2)
6p C = 3p N + 3p B + 6p H (3)
3p H = p C (4)
6p N - 6p B = 3p B - 3p N
fi 9p N = 9p B
fi pN = pB
6p C = 3p N + 3p N + 2p C
fi 4p C = 6p N
3
fi pC = 2
pN
pH = 1 3
3 2 ( )
pN = 1
p
2 N
(
(p N , p B , p C , p H ) = p N , p N , 32 p N , 21 p N )
Since these probabilities must add up to 1, this tells us that:
p N + p N + 32 p N + 21 p N = 1
fi 4p N = 1
fi pN = 1
4
So: (p N , p B , p C , p H ) = ( 1 1 3 1
, , ,
4 4 8 8 )
The long-term proportions are:
If we denote the state for markers on holiday by H , the transition graph for
this process looks like this:
0.1 0.6
A B
0.8 0.2
0.5 0.5
0.1 0.2
H
A B H
A È0.8 0.1 0.1˘
Í ˙
B Í 0.2 0.6 0.2˙
H ÍÎ0.5 0.5 0 ˙˚
The long-term probabilities for the 3 states can be found by finding the
stationary distribution, ie the set of probabilities p i that satisfy the matrix
equation p = p P with the additional condition Âpi = 1.
i
Written out in full, this matrix equation is:
p A - p B = 0.7p A - 0.4p B
fi 0.3p A = 0.6p B
fi p A = 2p B
(p A , p B , p H ) = (2p B , p B ,0.4p B )
2p B + p B + 0.4p B = 1
fi 3.4p = 1
B
fi pB =
1
3.4
= 0.2941 = ( ) 5
17
and (p A , p B , p H ) = ( 10 5 2
, ,
17 17 17 )
So the long-term proportions marking each subject are:
10
17
= 58.82% for Subject A
5
17
= 29.41% for Subject B
0.1 0.6
A B
0.8 0.2
1–x x
0.1 0.2
H
A B H
A È 0.8 0.1 0.1˘
Í ˙
B Í 0.2 0.6 0.2˙
H ÍÎ1 - x x 0 ˙˚
p B* = 0.8p B* + 0.2p B* + (1 - x )p H*
fi 0 = (1 - x )p H*
fi x =1
Yolanda will be covered by Company C for at least 5 years if she renews her
policy 4 times with Company C. This has probability:
0.4 4 = 0.0256
(ii) Probability Company A does not cover Zachary at the time of the fire
Company A will cover Zachary’s home at the time of the fire if he is insured
with Company A during the third year, which has probability pAA (2) .
So the probability that Company A does not cover his home at the time of
the fire is:
The long-term probabilities for the four states can be found by finding the
stationary distribution, ie the set of probabilities p i ≥ 0 that satisfy the matrix
equation p = p P with the additional condition Âpi = 1.
i
Written out in full, this matrix equation is:
We can subtract 0.4p B from each side of Equation (2) to make all the
coefficients on the RHS equal to 0.2. Using the fact that
p A + p B + p C + p D = 1 then gives:
p B - p C = 0.5p B - 0.2p C
fi 0.8p C = 0.5p B
5 5 1 5
fi pC = 8
pB = 8
¥ 3
= 24
5 31
p A = 1 - p B - p C - p D = 1 - 31 - 24
- 1
5
= 120
So: (p A , p B , p C , p D ) = ( 31 1 5 1
, , ,
120 3 24 5 ) = (0.2583, 0.3333, 0.2083, 0.2)
The long-term proportions are:
31
120
= 25.83% for Company A
1 = 33.33% for Company B
3
5
24
= 20.83% for Company C
1 = 20% for Company D.
5
AD B C
A È0.6 0.2 0.2 ˘
Í ˙
B Í0.3 0.6 0.1˙
C Í0.4 0.2 0.4 ˙
Í ˙
D ÍÎ0.6 0.2 0.2 ˙˚
AD B C
AD È0.6 0.2 0.2 ˘
Í ˙
B Í0.3 0.6 0.1˙
C ÍÎ0.4 0.2 0.4 ˙˚
Here the state labelled AD is for the new combined company (Addda).
(v) Comment
Irreducible?
Chain 1 is irreducible because we can cycle through all the states following
the circuit 1 Æ 2 Æ 3 Æ 1.
Chain 2 is irreducible because we can cycle through all the states following
the circuit 1 Æ 2 Æ 4 Æ 3 Æ 1 .
Periodic?
Chain 1 is aperiodic (not periodic) because we can return to any given state
in 2 or 3 or 4 or 5 etc steps, and these numbers are not restricted to a
multiple of some number greater than 1.
Chain 2 is periodic with period 2 because we can only return to any given
state in an even number of steps.
2
1–– 1 – 3
0% 20%
2
2
2
40%
2
1––
The transition matrix will be valid if the entries in each row add up to 1 (which
they do) and each entry lies in the range [0,1] .
We can see from the diagram in part (i) that, in general, it is possible to cycle
through the states in the order 0% Æ 20% Æ 40% Æ 0% . This means that
it is possible to move from any state i to any state j in a finite number of
steps, which is the definition of irreducible.
However, in the special case where b = 0 , the transition rates between the
states become zero and each state becomes an absorbing state. In this
case the chain is not irreducible.
The long-term probabilities for the 3 states can be found by finding the
stationary distribution, ie the set of probabilities p i that satisfy the matrix
equation p = p P with the additional condition Âpi = 1.
i
Written out in full, this matrix equation is:
23 143 13
11p 0 = 10p 20 + 12 p 20 = p
12 20
fi p0 = p
12 20
(p 0 , p 20 , p 40 ) = ( 13
p , p , 23 p
12 20 20 12 20 )
13 23
p
12 20
+ p 20 + 12 p 20 = 1
fi 4p 20 = 1
fi p 20 = 1 ( = 0.25)
4
and (p 0 , p 20 , p 40 ) = ( 13 1 23
, ,
48 4 48 ) = (0.271,0.25,0.479)
So the long-term proportions are:
13
48
= 27.1% with a 0% marginal tax rate
1 = 25% with a 20% marginal tax rate
4
23
48
= 47.9% with a 40% marginal tax rate.
Since Lucy currently pays 20% tax, the probabilities for Lucy’s tax rate for
2013 can be found from the middle row of entries in the matrix P 2 .
So the probabilities for Lucy’s tax rate in 2013 are 16.1% for 0% tax, 52% for
20% tax and 31.9% for 40% tax.
A Markov chain is a stochastic process that has a discrete time set and a
discrete state space and satisfies the Markov property, ie the future
probabilities depend only on the state currently occupied.
0 1 2 3 4
0 È0.7 0.3 0 0 0 ˘
Í ˙
1 Í 0.2 0.5 0.3 0 0 ˙
2 Í 0 0.2 0.5 0.3 0 ˙
Í ˙
3 Í 0 0 0.2 0.5 0.3 ˙
4 ÍÎ 0 0 0 0.2 0.8 ˙˚
The long-run probabilities for the five states are the entries in the stationary
distribution. These comprise the set of probabilities p i ≥ 0 that satisfy the
matrix equation p = p P , with the additional condition Âpi = 1.
i
0.3p 0 = 0.2p 1
0.3
fi p1 = p = 1.5p 0
0.2 0
fi 0.45p 0 = 0.2p 2
0.45
fi p2 = p = 2.25p 0
0.2 0
fi 0.675p 0 = 0.2p 3
0.675
fi p3 = p = 3.375p 0
0.2 0
fi 1.0125p 0 = 0.2p 4
1.0125
fi p4 = p 0 = 5.0625p 0
0.2
fi 13.1875p 0 = 1
1 5.0625
fi p0 = = 0.0758 and p 4 = = 0.3839
13.1875 13.1875
So the long-term probability that the rack is completely full or empty is:
Since this is greater than 35%, the city will increase the size of the rack.
(iii) Comment
If the size of the rack is increased, there will be more states for the process
to occupy, so we would expect the probabilities for each state to decrease.
However, we cannot be sure what will happen because the model makes
some big assumptions, such as assuming fixed values for the probabilities
0.3 and 0.2. In reality, these will vary depending on the time of day. Also,
increasing the size of the rack may encourage more people to come to work
by bike.
A B C
A È 0.6 0.2 0.2 ˘
Í ˙
B Í 0.25 0.6 0.15 ˙
C ÍÎ 0.1 0.4 0.5 ˙˚
The probabilities of being in each location on the third day are given by the
row for A in the matrix P 2 .
So the probabilities are: Atlantis = 43%, Beachy = 32% and Coral = 25%
(iii) Graph
0.6
0.5 A
0.4 B
B
0.3 A
B
C
0.2
C
0.1 B C
0 C
1 2 3 ... ∞
Day
We will need a 6-state model with the following states (where the bars
indicate the current state):
AB
AB
A ABC
AC
AC
Also, the entries in each row of the matrix must add up to 1. So:
A Ê 0.2 0.8 0 ˆ
B Á 0 0 1.0 ˜
Á ˜
C ÁË 0.6 0 0.4˜¯
For Path 1 in part (a) the sequence of states occupied at integer time points
is:
C A A A B C C A A B ?
All of the movements in this path have strictly positive probabilities, so this is
a valid sample path.
For Path 2 in part (b) the sequence of states occupied at integer time points
is:
A B B …
The movement from B to B has zero probability, so this is not a valid sample
path.
(i) Premium
The number of claims for each policyholder during the year has a
Poisson(0.35) distribution.
Since claims are paid up to a maximum of 3, the probabilities for the annual
number of claims from each policy are:
873.73
fi P= = £1,306.12
0.66895
So the full annual premium is £1,306.12 and a policyholder on the 40% level
will pay:
The transition graph for the protected NCD system looks like this (where pi
is the probability of making exactly i claims during the year).
p0 p0 p0 p0 + p1
1 – p0 1 – p0 – p1 1 – p0 – p1 1 – p0 – p1
p1 p1
(1 - 0.29531)p 0
p 15 = = 14.479p 0
0.04867
(1 - 0.24664)p 15 - 0.70469p 0
p 30 =
0.04867
0.75336(14.479p 0 ) - 0.70469p 0
=
0.04867
= 209.63p 0
0.70469p 30 0.70469(209.63p 0 )
p 40 = = = 3,035.1p 0
1 - 0.95133 1 - 0.95133
fi 3,260.2p 0 = 1
1
fi (p 0 , p 15 , p 30 , p 40 ) =
3,260.2
(1, 14.479, 209.63, 3035.1)
= (0.00031, 0.00444, 0.06430, 0.93096)
Assuming that the expected claim payments remain the same, this leads to
the new premium equation:
0.60766P = £873.73
873.73
fi P= = £1, 437.85
0.60766
So the full premium is £1,437.85 and a policyholder on the 40% level will
pay:
The protected system may encourage policyholders to stay with the same
insurer, as they start to build up a no claims history, which they do not want
to lose.
It should also discourage policyholders from claiming, which will reduce the
insurer’s costs.
However, it could also lead to bad publicity if some policyholders feel that
they have lost their NCD protection through no fault of their own (eg because
of an accident involving an uninsured driver).
FACTSHEET
Markov chains
A Markov process with a discrete time set and discrete state space is called
a Markov chain.
Using the Markov property and the law of total probability, the Chapman-
Kolmogorov equation for a time-inhomogeneous process is:
Here the transition probabilities depend only on the time lag t s and not on
the absolute values of s and t .
P t s P u s P t u
Stationary distributions
Irreducible means that each state can ultimately be reached starting from
any other, ie there exists at least one n 0 such that pij ( n ) 0 for all states
i and j .
A periodic chain with period d , is one where return to a given state is only
possible in a number of steps that is a multiple of d . NB: Return does not
need to be possible for all multiples of d .
Check this by finding the highest common factor of ‘return times’ for each
state. States in an irreducible chain have the same periodicity, so you only
need to check one state in this case.
n
This is defined as X n Yj where Y j are independent with the common
j 1
probability distribution:
n u n u
p 1 p if 0 n j i 2n and n j i is even
pij( n ) u
0 otherwise
Reflecting boundary: P X n 1 1 X n 0 1
Absorbing boundary: P X n 1 0 X n 0 1
P X 0X n 0 ,
n 1
Mixed boundary:
P X n 1 1X n 0 1
For a given driver, any period j is either accident free ( Y j 0 ) or gives rise
to exactly one accident ( Y j 1 ).
f y1 y 2 ... y n
P Yn 1 1 Y1 y1,Y2 y 2, ... , Yn y n
g n
f y1 y 2 ... y n
P Yn 1 0Y1 y1,Y2 y 2, ... , Yn y n 1
g n
Y1, Y2, ... , Yn , ... does not have the Markov property, but the cumulative
number of accidents experienced by the driver:
n
Xn Yj
j 1
nij
The best estimate of pij is pˆ ij .
ni
nijk to be Binomial Nij , p jk
The chi-square goodness-of-fit test is based on the test statistic:
nijk nij pˆ jk
2
2
m
i j k nij pˆ jk
Simulation
key assumptions
Markov property
– h qx +t = h m x + t + o(h ) as h Æ 0
= exp Ê - Ú m x + s ds ˆ
t
t px probability of staying alive
Ë 0 ¯
D d
m = or mˆ = formulae for MLE of m
V v
Ê m ˆ
m ~ N Á m, asymptotic distribution of MLE of m
Ë E [V ] ˜¯
specific points
this is a ‘warm-up’ model for general Markov jump processes
Write down f (d i , v i ) , the joint distribution of (Di ,Vi ) for one individual.
This uses the formula for t px .
The mean and variance of Di - mVi are then worked out. To do the mean:
The Core Reading mentions the variance as well, but it is unlikely you’d be
asked to prove this. (If you were, you’d be given appropriate guidance in the
question.)
m is an unbiased estimator of m , ie E ( m ) = m
Poisson model
parametric
– parameter is m (combined with E c as mE c )
key assumptions
likelihood function
c
e - mE x ( m E xc )d
L( m ) = P [D = d ] =
d!
D D
m = = formula for MLE of m
V E xc
E [ m ] = m , var[ ] properties of MLE of m
V E xc
exam questions
– numerical calculations
– writing down the likelihood function
– calculating the MLE
specific points
NOTES
NOTES
NOTES
NOTES
NOTES