Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
M347/F
Copyright
c 2018 The Open University
PART 1
• This part of the paper carries 50% of the total marks. (Allow some
time to check that your selections have been correctly entered on the
CME form.)
• You should attempt ALL the questions in this part of the
examination.
• You should note that for some of these questions you may be
required to select more than one answer from the options given. All
such questions include an instruction like ‘You should select TWO
options for this question’.
Question 2
Let X follow the beta distribution with pdf
f (x) = K(1 − x)3 on 0 < x < 1.
Choose the option that gives the value of K.
Question 3
Let X follow a binomial distribution with pmf
n x
p(x) = p (1 − p)n−x on x = 0, 1, . . . , n.
x
If E(X) = 4 and V (X) = 0.8 choose the option that gives the value of p.
Question 4
Suppose that, marginally, X follows an exponential, M(1), distribution
and that, conditionally, Y | X = x follows an (exponential) M(x)
distribution.
Select the option that gives the conditional distribution of X | Y = y.
Question 6
Suppose that X and Y follow a bivariate normal distribution with
μX = 1/2, μY = 1/2, σX = 1, σY = 1 and ρ = 1/2.
Choose from the following options the one statement that is TRUE.
Options for Question 6
A X − Y is normally distributed and V (X − Y ) = 1.
B X − Y is normally distributed and V (X − Y ) = 3/2.
C X − Y is normally distributed and Pr(X > Y ) = 0.
D X − Y is not normally distributed.
E X and Y are independent and Cov(X, Y ) = 0.
F X and Y are correlated and Cov(X, Y ) = 1.
Question 7
Consider the estimator of the population mean μ that has the form
1√
n
n + Xi
μ̃ = 2 √ i=1 .
n+ n
Choose from the following options the Cramér-Rao lower bound for the
variance of any unbiased estimator of θ.
Question 9
Suppose that independent observations x1 , x2 , . . . , xn are available from
the gamma distribution with parameters 3 and b, which has pdf
f (x) = 12 b3 x2 e−bx .
The log-likelihood is
n
l(b) = −n log 2 + 3n log b + 2 log xi − bnx
i=1
Question 11
Let X be a random variable that has a symmetric distribution with
mean 0 and upper quartile q3/4 . Use Chebyshev’s inequality to choose
from the following options the correct inequality for the variance V (X)
relative to the upper quartile q3/4 .
Question 13
Let X be an observation from the distribution with pmf
f (x|θ) = (x − 1)θ 2 (1 − θ)x−2 , on x = 2, 3, . . .
where 0 < θ < 1.
A prior is required for θ which represents a lack of any idea about the
value of θ.
Choose from the following options the most suitable choice of prior
for θ.
Question 14
The number of phone calls received by a specific help-line in a day is
assumed to follow a Poisson distribution with mean λ. The prior
knowledge about the value of λ corresponds to the gamma distribution
1 2 −λ/3
f (λ) = λe , on λ > 0.
54
In the next 4 days the help-line received 3, 7, 6 and 10 phone calls,
respectively.
Choose from the following options the posterior distribution of λ.
Question 16
Consider the following events.
• A : X1 , X2 , . . . , Xn are independent and come from a geometric
distribution with pmf
f (x|θ1 ) = θ1 (1 − θ1 )x , on x = 0, 1, . . .
with 0 < θ1 < 1;
• Ac : X1 , X2 , . . . , Xn are independent and come from a geometric
distribution with pmf
f (x|θ2 ) = θ2 (1 − θ2 )x , on x = 0, 1, . . .
with 0 < θ2 < 1.
Assuming that you have observed a sample mean x = 3, choose from
the following options the correct statement about the Bayes factor for
A against Ac when θ1 = 1/3 and θ2 = 2/3.
Question 18
The matrix
a b
P= 1 3
4 4
Question 20
Where appropriate, linear regression with one explanatory variable can
be performed ‘through the origin’, that is, with the intercept α fixed
equal to zero. The log-likelihood in this reduced model is
S0 (β)
l(β, σ) = constant − n log σ −
2σ 2
where
n
S0 (β) = (yi − βxi )2 .
i=1
Solve the equation dS0 (β)/dβ = 0 to provide the candidate value β for
the value of β that minimises S0 (β). (It can be confirmed that β thus
found is the MLE for β but you need not do so.)
Choose the correct value of β from the following options.
n n
(x − x)(yi − y) x (y − y)
E n i
i=1
F ni i 2
i=1
i=1 (xi − x)
2
i=1 xi
Question 22
In the multiple regression model, the covariance matrix of β is
Choose from the
σ 2 (X T X)−1 . The fitted values are given by Y = X β.
following options the covariance matrix of Y .
Question 23
Let Y be the sum of three independent variables with distributions
M(λ), so Y ∼ Gamma(3, λ). This distribution is a member of the
exponential dispersion family with a(λ) = −λ and mean μ = 3/λ.
Select the formula for its canonical link.
Question 26
Let X be a continuous random variable following the distribution with
cdf
9
FX (x) = 1 − 2 on x > 3.
x
(a) What is the value of Pr(X > 9)? [2]
(b) Find the pdf, f (x). [2]
(c) Show that the quantile function can be written
3
Q(α) = √ . [2]
1−α
(e) Define Y = 1/X. Find the cdf of Y . (You may take it for granted
that the support of Y is 0 < y < 1/3). [2]
Question 27
Let (X, Y ) follow the bivariate distribution with pdf
1
f (x, y) = on 0 < x < y, 0 < y < 1.
y
(a) Show that the Y -marginal of this distribution is the U(0, 1)
distribution, that is, the uniform distribution with parameters 0
and 1. [2]
(b) Show that, for each value of 0 < y < 1, the conditional density of
X|Y = y is
1
fX|Y (x|y) = on 0 < x < y. [2]
y
(c) Name the distribution whose pdf is obtained in part (b), and state
its parameters. Hence, give the formula for E(X|Y = y). [2]
(d) Use your answer in part (c) to show that E(X) = 1/4. [2]
(e) Find E(XY ). [2]
(b) Find
(β) and hence show that the candidate MLE is β = r. [2]
(c) Confirm that β = r is indeed the MLE of β. [3]
= β and
(d) It turns out that for this Rayleigh distribution, E(β)
= β 2 /n. What is the formula for the mean squared error of β
V (β)
and
as an estimator of β? How does this formula relate to V (β),
why? [2]
Question 29
Let independent positive random variables X1 , X2 , . . . , Xn be modelled
by a gamma distribution with parameters a = 2 (known) and θ > 0
(unknown), so that for i = 1, 2, . . . , n,
f (xi | θ) = θ2 xi exp(−θxi ) on xi > 0.
(a) Show that L(θ), the likelihood for θ based on observed data
x = (x1 , x2 , . . . , xn )T , can be written
L(θ) ∝ θ2n exp(−θnx). [2]
(b) Suppose that a Gamma(a, b) prior is specified for θ. Show that the
posterior f (θ|x) can be written
f (θ|x) ∝ θa+2n−1 exp{−(b + nx)θ}. [3]
(c) Suppose that the candidate value generated from the distribution
with density q(λ2 |λ1 ) is λ∗ = 0.1 and that the acceptance
probability in part (b) turns out to be 0.492. Suppose also that u is
simulated from U(0, 1) so that u = 0.477. What is the value of λ2 ?
From what distribution will the candidate value λ∗∗ for λ3 be
generated? [2]
(d) After several thousand sampled values of λ have been generated
according to the above algorithm, an autocorrelation function is
calculated. It takes the value 1 when the lag (the time between
observations) is zero, 0.4 when the lag is one, 0.15 when the lag is
two, and 0 for all higher lags. It is decided to thin the series of
values obtained. Say what is meant by thinning in this context,
why one would do it, and give an appropriate value for how much
to thin the sample. [3]
(e) When all full conditionals of the posterior distribution of a
multidimensional vector of parameters are available, it is
appropriate to use a method that is a special case of the
Metropolis–Hastings algorithm. Explain what is meant by a “full
conditional” and name that special case method. [2]
(c) The trace of a square matrix is the sum of its diagonal elements.
For example, tr(I n ) = n. Also, tr(AB) = tr(BA) for any A and
B whose product is a square matrix. Use these facts to evaluate
tr(H) where the hat matrix
H = X(X T X)−1 X T
in terms of the observed values y through
gives the fitted values y
= Hy.
y
The expectation of the residual sum of squares, R, in the multiple
regression model can be written as
E(R) = σ 2 {tr(I n ) − tr(H)}.
Hence give an unbiased estimator for σ 2 in terms of R, n and d. [3]
Question 32
Suppose the random variable Y follows the geometric distribution with
parameter p ∈ (0, 1):
f (y|p) = p(1 − p)y−1 on y = 1, 2, . . . .
Then Y has a distribution belonging to the exponential dispersion
family (EDF) with θ = p and dispersion parameter φ = 1. (You can
therefore set d(φ) = 1 below.)
(a) Identify the functions a(p) and b(p) in the standard formula for the
pmf of a discrete member of the EDF. [4]
(b) Use general properties of the EDF to show that E(Y ) = 1/p and
V (Y ) = (1 − p)/p2 . [5]
(c) Identify the canonical link for the geometric distribution. [1]
1
Bivariate and multivariate normal distribution
n
2
Cov(Z, W ) = σ j mj .
j=1
2
Gamma and beta functions
• For a > 0,
∞
Γ(a) = xa−1 e−x dx,
0
√
Γ(a + 1) = a Γ(a), Γ( 12 ) = π,
and, for integer n = 1, 2, . . ., Γ(n) = (n − 1)!.
• For a, b > 0,
1
B(a, b) = xa−1 (1 − x)b−1 dx,
0
Γ(a) Γ(b)
B(a, b) = .
Γ(a + b)
Some asymptotics
p
• Xn converges in probability to X, denoted Xn −→ X, if
P (|Xn − X| > ε) → 0 as n → ∞, for any ε > 0.
ms
• Xn converges in mean square to X, denoted Xn −→ X, if
E{(Xn − X)2 } → 0 as n → ∞.
D
• Xn converges in distribution to X, denoted Xn −→ X, if
Fn (x) → F (x) as n → ∞,
for all values x where F (x) does not jump.
• Let θn be the MLE of θ based on a random sample of size n.
Subject to some regularity conditions,
√ 1
n(
D
θn − θ) −→ W ∼ N 0, .
i(θ)
3
Some important theorems
Bayes’ Theorem
For conditional and marginal distributions,
fX|Y (x|y) fY (y)
fY |X (y|x) = ;
fX (x)
for prior and posterior pdfs or pmfs,
f (x|θ) f (θ)
f (θ|x) = .
f (x)
Neyman–Pearson Lemma
When performing a hypothesis test between H0 : θ = θ0 and
H1 : θ = θ1 , the likelihood ratio test with rejection region of the form
R = {x : L(θ1 )/L(θ0 ) > c} is the most powerful test with a given size.
Gauss–Markov Theorem
is any solution of the normal
If kT θ is an estimable function and θ
T
equations, then k θ has minimum variance in the class of linear
unbiased estimators of kT θ.
4
Test statistics
Let
θ be the MLE of θ, and the log-likelihood. For testing H0 : θ = θ0
against H1 : θ = θ0 , four test statistics based on likelihood theory are:
Inequalities
5
Bayesian and Markovian miscellany
• The Bayes factor for event A against event Ac is given by
B(x, A) = f (x|A)/f (x|Ac ).
• The equilibrium distribution in a discrete Markov chain satisfies
πT = πT P .
• The detailed balance equation for a continuous Markov chain
with transition kernel k(xt+1 |xt ) and equilibrium density π is, for
all x, y ∈ S,
k(y|x) π(x) = k(x|y) π(y);
and for a discrete Markov chain, it is, for all i, j ∈ S,
P (Xt+1 = j | Xt = i) πi = P (Xt+1 = i | Xt = j) πj .
• The acceptance probability in the Metropolis–Hastings algorithm
is
∗ q(xt |x∗ ) f (x∗ )
α(x |xt ) = min ,1 .
q(x∗ |xt ) f (xt )
6
• In the general linear model, the normal equations are
(AT A)θ = AT y.
• In multiple regression, the multivariate normal-gamma prior is a
conjugate prior for θ = (βT τ)T , so that
β, τ ∼ MNgamma(a, B, c, d), with
β|τ ∼ N(a, B/τ) and τ ∼ Gamma(c, d).
The posterior distribution is given by
θ|y ∼ MNgamma(a1 , B 1 , c1 , d1 ),
where
−1
B 1 = B −1 + X T X , a1 = B 1 (B −1 a + X T y),
n 1
c1 = c + , d1 = d + (y T y + aT B −1 a − aT1 B −11 a1 ).
2 2