var_bayesian_analysis
var_bayesian_analysis
Frank Schorfheide
Department of Economics, University of Pennsylvania
Frank Schorfheide, University of Pennsylvania: Bayesian Methods 2
Preliminaries
• Then
³ ´
0 0 −1
β|Σ, X X ∼ N β̂, Σ ⊗ (X X) . (2)
C is the lower triangular Cholesky decomposition matrix. Then let Z = µ+CN (0, I).
Frank Schorfheide, University of Pennsylvania: Bayesian Methods 3
Preliminaries
• The multivariate version of the inverted Gamma distribution is called Wishart Distri-
bution.
• Let Σ be a n×n positive definite random matrix. Σ has the inverted Wishart IW (S, ν)
Preliminaries
• Recall:
½ ¾ ½ ¾
1 1
p(Y |Φ, Σ, Y0) ∝ |Σ|−T /2 exp − tr[Σ−1S] exp − tr[Σ−1(Φ − Φ̂)0X 0X(Φ − Φ̂)]
2 2
Preliminaries
• Hence,
• Define
0 0
Φ̂∗ = (X ∗ X ∗)−1X ∗ Y ∗
S ∗ = (Y ∗ − X ∗Φ̂∗)0(Y ∗ − X ∗Φ̂∗).
• It can be verified that the prior p(Φ, Σ|Y ∗) is of the Inverted Wishart-Normal IW −N
form
µ ¶
Σ ∼ IW S ∗, T ∗ − k (6)
µ ¶
0
Φ|Σ ∼ N Φ∗, Σ ⊗ (X ∗ X ∗)−1 . (7)
Frank Schorfheide, University of Pennsylvania: Bayesian Methods 8
nk ∗
∗0 ∗ − n2 ∗ − T 2−k
c∗ = (2π) |X 2 X | |S | (8)
n
Y
n(T ∗ −k) n(n−1)
2 2 π 4 Γ[(T ∗ − k + 1 − i)/2],
i=1
• The implementation of priors through dummy variables is often called mixed estimation
• Notice that
• Define:
0 0
Φ̃ = (X ∗ X ∗ + X 0X)−1(X ∗ Y ∗ + X 0Y ) (10)
·
1 0
Σ̃u = ∗ Y ∗ Y ∗ + Y 0Y )
T +T
¸
∗0 ∗0 ∗0
−(X Y ∗ + X 0Y )0(X X ∗ + X 0X)−1(X Y ∗ + X 0Y ) . (11)
Frank Schorfheide, University of Pennsylvania: Bayesian Methods 10
• Since prior and likelihood function are conjugate, it is straightforward to show, that
the posterior distribution of Φ and Σ is also of the Inverted Wishart – Normal form:
µ ¶
Σ|Y ∼ IW (T ∗ + T )Σ̃u, T ∗ + T − k (12)
µ ¶
0
Φ|Σ, Y ∼ N Φ̃, Σ ⊗ (X ∗ X ∗ + X 0X)−1 . (13)
• Suppose that we are using a prior constructed from dummy observations. Then the
• The integrals in the numerator and denominator are given by the appropriate modifi-
where
Φ̂ = (X 0X)−1X 0Y
S = (Y − X Φ̂)0(Y − X Φ̂).
Frank Schorfheide, University of Pennsylvania: Bayesian Methods 12
• Minnesota Prior
Minnesota Prior
• Reference: Doan, Litterman, and Sims (1984). The version below is described in the
Minnesota Prior
or in matrix form
Y = XΦ + U. (19)
• Based on a short pre-sample Y0 (typically the observations used to initialized the lags
Minnesota Prior
– τ is the overall tightness of the prior. Large values imply a small prior covariance
matrix.
– d: the variance for the coefficients of lag h is scaled down by the factor l−2d.
– w: determines the weight for the prior on Σu. Suppose that Zi = N (0, σ 2). Then
P
an estimator for σ 2 is σ̂ 2 = w1 wi=1 Zi2. The larger w, the more informative the
estimator, and in the context of the VAR, the tighter the prior.
Minnesota Prior
Minnesota Prior
Minnesota Prior
• Co-persistence prior dummy observations, reflecting the belief that when data on all
y’s are stable at their initial levels, thy will tend to persist at that level:
· ¸ · ¸
0
λȳ1 λȳ2 = λȳ1 λȳ2 λȳ1 λȳ2 λ Φ + u
• Own-persistence prior dummy observations, reflecting the belief that when yi has been
stable at its initial level, it will tend to persist at that level, regardless of the value of
other variables:
µȳ1 0 µȳ1 0 µȳ1 0 0 0
= Φ + u
0 µȳ2 0 µȳ2 0 µȳ2 0
Frank Schorfheide, University of Pennsylvania: Bayesian Methods 19
• Training Sample Prior: replace dummy observations by actual observations from a pre-
or training sample.
• DSGE Model Prior: use artificial observations generated by a DSGE model. Details