Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Model
yi = β1 + β2 xi 2 + ... + βk xik + εi .
y = X β + ε.
β=V V 1
β + X 0X b
β
ν = ν+N
and s 2 is de…ned implicitly through
0 h i 1
1
νs 2 = νs 2 + νs 2 + b
β β V + X 0X b
β β .
β jy t β, s 2 V , ν ,
E ( β jy ) = β
νs 2
var ( βjy ) = V.
ν 2
Intuition: Posterior mean and variance are weighted average of
information in the prior and the data.
Mj : yj = Xj β(j ) + εj
M2 is unrestricted model
y = β1 + β2 x2 + β3 x3 + ε
y x3 = β1 + β2 x2 + ε
p (yj jMj ) = cj νj s 2j 2
jV j j
Posterior odds ratio depends on the prior odds ratio and contains
rewards for model …t, coherency between prior and data information
and parsimony.
Want to predict:
y = X β+ε
Remember, prediction is based on:
Z Z
p (y jy ) = p (y jy , β, h ) p ( β, h jy )d βdh.
p a βj b jy = 0.95.
Before we had conjugate prior where p ( βjh ) was Normal density and
p (h ) Gamma density.
Now use similar prior, but assume prior independence between β and
h.
p ( β, h ) = p ( β) p (h ) with p ( β) being Normal and p (h ) being
Gamma:
β N β, V
and
2
h G (s , ν)
Key di¤erence: now V is now the prior covariance matrix of β, with
conjugate prior we had var ( βjh ) = h 1 V .
β=V V 1
β + hX 0 y
where
ν = N +ν
and
(y X β)0 (y X β) + νs 2
s2 =
ν
Econometrician is interested in p ( β, h jy ) (or p ( βjy )), NOT the
posterior conditionals, p ( βjy , h ) and p (h jy , β).
Since p ( β, h jy ) 6= p ( βjy , h ) p (h jy , β), the conditional posteriors do
not directly tell us about p ( β, h jy ).
But, there is a posterior simulator, called the Gibbs sampler, which
uses conditional posteriors to produce random draws, β(s ) and h(s ) for
s = 1, .., S, which can be averaged to produce estimates of posterior
properties just as with Monte Carlo integration.
You can continue this reasoning inde…nitely producing β(s ) , h(s ) for
s = 1, .., S
y = X β + ε.
Before assumed ε was N (0N , h 1I
N ).
Many other models involve
ε N (0N , h 1
Ω)
y † = X † β + ε†
p ( β, h, Ω) = p ( β) p (h ) p (Ω)
where:
β N β, V
and
2
h G (s , ν)
βjy , h, Ω N β, V
where
1
V = V 1
+ hX 0 Ω 1
X
β=V V 1
β + hX 0 Ω 1
Xb
β (Ω)
h jy , β, Ω G (s 2
, ν ),
where b
β (Ω) is the GLS estimator
ν = N +ν
and
(y X β)0 Ω 1
(y X β) + νs 2
s2 =
ν
p (Ωjy , β, h ) ∝
1
p (Ω) jΩj 2 exp h
2 (y X β)0 Ω 1
(y X β)
will converge to E [g (y ) jy ].
The following strategy provides draws of y .
For every β(s ) , h(s ) , Ω(s ) from Gibbs sampler, take a draw (or several)
of y (s ) from p y j β(s ) , h(s ) , Ω(s ) (a Normal density)
We now have draws β(s ) , h(s ) , Ω(s ) and y (s ) for s = 1, .., S which we
can use for posterior or predictive inference.
Why are these the correct draws? Use rules of conditional probability
(see pages 72-73 of textbook for details).
We begin by eliciting p (ω ).
Work with error precisions rather than variances and, hence, we de…ne
0
λ (λ1 , λ2 , .., λN )0 ω 1 1 , ω 2 1 , .., ω N 1 .
Consider the following prior for λ:
N
p (λ) = ∏ fG (λi j1, νλ ) (**)
i =1
Why should the λi s be i.i.d. draws from the Gamma distribution with
mean 1.0?
Can prove this model is exactly the same as the linear regression
model with i.i.d. Student-t errors with νλ degrees of freedom
(Bayesian Econometric Methods Exercise 15.1).
In other words, if we had begun by assuming:
1
p (εi ) = ft εi j0, h , νλ
Note: we now have model with more ‡exible error distribution, but we
are still our familiar Normal linear regression model framework.
Note: a popular way of making models/distributions more ‡exible is
through: mixture of Normals distributions.
Our treatment here is an example of a scale mixture of Normals.
If νλ is unknown, need a prior p (νλ ).
Note that now the prior for λ is speci…ed in two steps, the …rst being
(**), the other being p (νλ ).
Alternatively, the prior for λ can be written as p (λjνλ ) p (νλ ).
Priors written in two (or more) steps in this way are referred to as
hierarchical priors.
νλ + 1
p (λi jy , β, h, νλ ) = fG λi j , νλ + 1
hε2i + νλ
α θ (s 1)
,θ =
p (θ =θ jy )q (θ ;θ =θ (s 1 ) )
min ,1
p ( θ = θ (s 1)
jy )q (θ (s 1 ) ;θ =θ )
θ = θ (s 1)
+w
q θ (s 1)
; θ = fN (θ jθ (s 1)
, Σ ).