Processing math: 23%
+ - 0:00:00
Notes for current slide
Notes for next slide

Lecture 5

Dynamics theory review

Ivan Rudik

AEM 7130

1 / 32

Roadmap

  1. Review Markov chains and dynamic models
  2. Review theory for numerical methods
2 / 32

Building a dynamic economic model

We need 5 things for a dynamic economic model

3 / 32

Building a dynamic economic model

We need 5 things for a dynamic economic model

  1. Controls: what variables are we optimizing, what decisions do the economic agents make?
3 / 32

Building a dynamic economic model

We need 5 things for a dynamic economic model

  1. Controls: what variables are we optimizing, what decisions do the economic agents make?

  2. States: What are the variables that change over time and interact with the controls?

3 / 32

Building a dynamic economic model

We need 5 things for a dynamic economic model

  1. Controls: what variables are we optimizing, what decisions do the economic agents make?

  2. States: What are the variables that change over time and interact with the controls?

  3. Payoff: What is the single-period payoff function? What's our reward?

3 / 32

Building a dynamic economic model

We need 5 things for a dynamic economic model

  1. Controls: what variables are we optimizing, what decisions do the economic agents make?

  2. States: What are the variables that change over time and interact with the controls?

  3. Payoff: What is the single-period payoff function? What's our reward?

  4. Transition equations: How do the state variables evolve over time?

3 / 32

Building a dynamic economic model

We need 5 things for a dynamic economic model

  1. Controls: what variables are we optimizing, what decisions do the economic agents make?

  2. States: What are the variables that change over time and interact with the controls?

  3. Payoff: What is the single-period payoff function? What's our reward?

  4. Transition equations: How do the state variables evolve over time?

  5. Planning horizon: When does our problem terminate? Never? 100 years?

3 / 32

Two types of solutions

Dynamic problems can generally be solved in two ways

4 / 32

Two types of solutions

Dynamic problems can generally be solved in two ways

Open-loop: treat the model as a sequence of static optimization problems solved simultaneously

4 / 32

Two types of solutions

Dynamic problems can generally be solved in two ways

Open-loop: treat the model as a sequence of static optimization problems solved simultaneously

  • Transitions act as constraints
  • Ends up being just a potentially giant (but simple) non-linear optimization problem
  • Drawback: solutions will be just a function of time so we can't introduce uncertainty, strategic behavior, etc
4 / 32

Two types of solutions

Feedback: treat the model as a single-period optimization problem with the immediate payoff and the continuation value

5 / 32

Two types of solutions

Feedback: treat the model as a single-period optimization problem with the immediate payoff and the continuation value

  • Yields a solution that is a function of states
  • Permits uncertainty, game structures
  • Drawback: need to solve for the continuation value function
5 / 32

Markov chains

Dynamic models in economic models are typically Markovian

6 / 32

Markov chains

Dynamic models in economic models are typically Markovian

A stochastic process {xt} is said to have the Markov property if for all k1 and all t Prob(xt+1|xt,xt1,...,xtk)=Prob(xt+1|xt)

6 / 32

Markov chains

Dynamic models in economic models are typically Markovian

A stochastic process {xt} is said to have the Markov property if for all k1 and all t Prob(xt+1|xt,xt1,...,xtk)=Prob(xt+1|xt)

The distribution of the next vector in the sequence (i.e. the distribution of next period's state) is a function of only the current vector (state)

6 / 32

Markov chains

Dynamic models in economic models are typically Markovian

A stochastic process {xt} is said to have the Markov property if for all k1 and all t Prob(xt+1|xt,xt1,...,xtk)=Prob(xt+1|xt)

The distribution of the next vector in the sequence (i.e. the distribution of next period's state) is a function of only the current vector (state)

The Markov property is necessary for the feedback representation

6 / 32

Markov chains

We characterize stochastic state transitions with Markov chains

7 / 32

Markov chains

We characterize stochastic state transitions with Markov chains

A Markov chain is characterized by:

  1. n-dimensional state space with vectors ei, i=1,...,n where ei is an n×1 unit vector whose ith entry is 1 and all others are 0
  2. An n×n transition matrix P which captures the probability of transitioning from one point of the state space to another point of the state space next period
  3. n×1 vector π0 whose ith value is the probability of being in state i at time 0: π0i=Prob(x0=ei)
7 / 32

Markov chains

P is given by Pij=Prob(xt+1=ej|xt=ei)

8 / 32

Markov chains

P is given by Pij=Prob(xt+1=ej|xt=ei)

We need one assumption:

  • For i=1,..,n, nj=1Pij=1 and π0 satisfies: ni=1π0i=1
8 / 32

Markov chains

Nice property of Markov chains:
We can use P to determine the probability of moving to another state in two periods by P2 since Prob(xt+2=ej|xt=ei)=nh=1Prob(xt+2=ej|xt+1=eh)Prob(xt+1=eh|xt=ei)=nh=1PihPhj=P2ij

9 / 32

Markov chains

Nice property of Markov chains:
We can use P to determine the probability of moving to another state in two periods by P2 since Prob(xt+2=ej|xt=ei)=nh=1Prob(xt+2=ej|xt+1=eh)Prob(xt+1=eh|xt=ei)=nh=1PihPhj=P2ij

iterate on this to show that Prob(xt+k=ej|xt=ei)=Pkij

9 / 32

Dynamic programming

Start with a general sequential problem to set up the basic recursive/feedback dynamic optimization problem

10 / 32

Dynamic programming

Start with a general sequential problem to set up the basic recursive/feedback dynamic optimization problem

Let β(0,1), the economic agent selects a sequence of controls, {ut}t=0 to maximize t=0βtr(xt,ut) subject to xt+1=g(xt,ut) and with x0 given

10 / 32

Dynamic programming

Assume r is concave, continuously differentiable, and the state space is convex and compact

11 / 32

Dynamic programming

Assume r is concave, continuously differentiable, and the state space is convex and compact

We want to recover a policy function h which maps the current state xt into the current control ut, such that the sequence {us}s=0 generated by iterating ut=h(xt)xt+1=g(xt,ut), starting from x0, solves our original optimization problem

11 / 32

Value functions

Consider a function V(x), the continuation value function where V(x0)=max subject to the transition equation: x_{t+1} = g(x_t,u_t)

The value function defines the maximum value of our original problem as a function of the state

12 / 32

Value functions

Suppose we know V(x), then we can solve for the policy function h by solving for each x \in X \max_u r(x,u) + \beta V(x') where x' = g(x,u) and primes on state variables indicate next period

13 / 32

Value functions

Suppose we know V(x), then we can solve for the policy function h by solving for each x \in X \max_u r(x,u) + \beta V(x') where x' = g(x,u) and primes on state variables indicate next period

Conditional on having V(x) we can solve our dynamic programming problem

13 / 32

Value functions

Suppose we know V(x), then we can solve for the policy function h by solving for each x \in X \max_u r(x,u) + \beta V(x') where x' = g(x,u) and primes on state variables indicate next period

Conditional on having V(x) we can solve our dynamic programming problem

Instead of solving for an infinite dimensional set of policies, we instead find the V(x) and h that solves the continuum maximization problems, where there is a unique maximization problem for each x

13 / 32

Value functions

Suppose we know V(x), then we can solve for the policy function h by solving for each x \in X \max_u r(x,u) + \beta V(x') where x' = g(x,u) and primes on state variables indicate next period

Conditional on having V(x) we can solve our dynamic programming problem

Instead of solving for an infinite dimensional set of policies, we instead find the V(x) and h that solves the continuum maximization problems, where there is a unique maximization problem for each x

This is often easier

13 / 32

Bellman equations

Issue: How do we know V(x) when it depends on future (optimized) actions?

14 / 32

Bellman equations

Issue: How do we know V(x) when it depends on future (optimized) actions?

Define the Bellman equation V(x) = \max_u r(x,u) + \beta V[g(x,u)]

14 / 32

Bellman equations

Issue: How do we know V(x) when it depends on future (optimized) actions?

Define the Bellman equation V(x) = \max_u r(x,u) + \beta V[g(x,u)]

h(x) maximizes the right hand side of the Bellman

14 / 32

Bellman equations

The policy function satisfies V(x) = r[x,h(x)] + \beta V\{g[x,h(x)]\}

15 / 32

Bellman equations

The policy function satisfies V(x) = r[x,h(x)] + \beta V\{g[x,h(x)]\}

Solving the problem yields a solution that is a function, V(x)

15 / 32

Bellman equations

The policy function satisfies V(x) = r[x,h(x)] + \beta V\{g[x,h(x)]\}

Solving the problem yields a solution that is a function, V(x)

This is a recursive problem since it maps itself into a scalar value, can be hard to think about at first

15 / 32

Bellman equations

The policy function satisfies V(x) = r[x,h(x)] + \beta V\{g[x,h(x)]\}

Solving the problem yields a solution that is a function, V(x)

This is a recursive problem since it maps itself into a scalar value, can be hard to think about at first

One of the workhorse solution methods exploits this recursion and
contraction mapping properties of the Bellman operator to solve for V(x)

15 / 32

Solution properties

Under standard assuptions we have that

  1. The solution to the Bellman equation, V(x), is strictly concave
  2. The solution is approached in the limit as j \rightarrow \infty by iterations on:
    V_{j+1}(x) = \max_{u} r(x,u) + \beta V_j(x'), given any bounded and continuous V_0 and our transition equation
  3. There exists a unique and time-invariant optimal policy function u_t = h(x_t)
    where h maximizes the right hand side of the Bellman
  4. The value function V(x) is differentiable
16 / 32

Euler equations

Euler equations are dynamic efficiency conditions: they equalize the marginal effects of an optimal policy over time

E.g: set the current marginal benefit, energy from burning fossil fuels, with the future marginal cost, global warming

17 / 32

Euler equations

Euler equations are dynamic efficiency conditions: they equalize the marginal effects of an optimal policy over time

E.g: set the current marginal benefit, energy from burning fossil fuels, with the future marginal cost, global warming

  1. We have a stock of capital K_t that depreciates at rate \delta \in (0,1)
  2. We can invest to increase our future capital I_t at cost c(I_t) and effectiveness \gamma \in (0,1]
  3. Per-period payoff U(Y_t) from consuming output Y_t = f(K_t) = K_t
  4. Discount factor is \beta \in (0,1)
17 / 32

Euler equations

The Bellman equation is \begin{align} V(K_t) &= \max_{I_t} \left\{ u(K_t) - c(I_t) + \beta V(K_{t+1}) \right\} \notag \\ &\text{subject to: } \,\,\,\, K_{t+1} = (1 - \delta) K_t + \gamma I \notag \end{align}

18 / 32

Euler equations

The Bellman equation is \begin{align} V(K_t) &= \max_{I_t} \left\{ u(K_t) - c(I_t) + \beta V(K_{t+1}) \right\} \notag \\ &\text{subject to: } \,\,\,\, K_{t+1} = (1 - \delta) K_t + \gamma I \notag \end{align}

The FOC with respect to investment is c_I(I_t) = \beta \, \gamma \, V_K(K_{t+1})

18 / 32

Euler equations

The Bellman equation is \begin{align} V(K_t) &= \max_{I_t} \left\{ u(K_t) - c(I_t) + \beta V(K_{t+1}) \right\} \notag \\ &\text{subject to: } \,\,\,\, K_{t+1} = (1 - \delta) K_t + \gamma I \notag \end{align}

The FOC with respect to investment is c_I(I_t) = \beta \, \gamma \, V_K(K_{t+1})

Envelope theorem gives us V_K(K_t) = u_K(K_t) + \beta \, \delta \, V_K(K_{t+1})

18 / 32

Euler equations

The FOC with respect to investment is c_I(I_t) = \beta \, \gamma \, V_K(K_{t+1})

Envelope theorem gives us V_K(K_t) = u_K(K_t) + \beta \, \delta \, V_K(K_{t+1})

19 / 32

Euler equations

The FOC with respect to investment is c_I(I_t) = \beta \, \gamma \, V_K(K_{t+1})

Envelope theorem gives us V_K(K_t) = u_K(K_t) + \beta \, \delta \, V_K(K_{t+1})

Advance both by one period since they must hold for all t

\begin{gather} c_I(I_{t+1}) = \beta \, \gamma \, V_K(K_{t+2}) \notag\\ V_K(K_{t+1}) = u_K(K_{t+1}) + \beta \, \delta \, V_K(K_{t+2}) \notag \end{gather}

19 / 32

Euler equations

Substitute the time t and time t+1 FOCs into our time t+1 envelope condition \frac{c'(I_t)}{\beta \, \gamma} = u'(K_{t+1}) + \beta \, \delta \frac{c'(I_{t+1})}{\beta \, \gamma} \Rightarrow c'(I_t) = \beta \left[ \gamma \, u'(K_{t+1}) + \delta \, c'(I_{t+1}) \right]

20 / 32

Euler equations

Substitute the time t and time t+1 FOCs into our time t+1 envelope condition \frac{c'(I_t)}{\beta \, \gamma} = u'(K_{t+1}) + \beta \, \delta \frac{c'(I_{t+1})}{\beta \, \gamma} \Rightarrow c'(I_t) = \beta \left[ \gamma \, u'(K_{t+1}) + \delta \, c'(I_{t+1}) \right]

LHS is marginal cost of investment, RHS is marginal benefit of investment along an optimal path

20 / 32

Euler equations

\Rightarrow c'(I_t) = \beta \left[ \gamma \, u'(K_{t+1}) + \delta \, c'(I_{t+1}) \right]

LHS: marginal cost of investment
RHS: marginal benefit of higher utility from more future output, and lower future investment cost because of higher capital stock

21 / 32

Euler equations: no-arbitrage

Euler equations are no-arbitrage conditions

Suppose we're on the optimal capital path and want to deviate by cutting back investment

22 / 32

Euler equations: no-arbitrage

Euler equations are no-arbitrage conditions

Suppose we're on the optimal capital path and want to deviate by cutting back investment

Yields a marginal benefit today of saving us some investment cost

22 / 32

Euler equations: no-arbitrage

Euler equations are no-arbitrage conditions

Suppose we're on the optimal capital path and want to deviate by cutting back investment

Yields a marginal benefit today of saving us some investment cost

There are two costs associated with it:

22 / 32

Euler equations: no-arbitrage

Euler equations are no-arbitrage conditions

Suppose we're on the optimal capital path and want to deviate by cutting back investment

Yields a marginal benefit today of saving us some investment cost

There are two costs associated with it:

  1. Lower utility tomorrow because we will have a smaller capital stock

  2. Greater investment cost tomorrow to return to the optimal capital trajectory

22 / 32

Euler equations: no-arbitrage

If this deviation (or deviating by investing more today) were profitable, we would do it

\rightarrow the optimal policy must have zero additional profit opportunities: this is what the Euler equation defines

23 / 32

Basic theory

Here we finish up the basic theory pieces we need

We will focus on deterministic problems but this easily ports to stochastic problems

24 / 32

Basic theory

Here we finish up the basic theory pieces we need

We will focus on deterministic problems but this easily ports to stochastic problems

Consider an infinite horizon problem for an economic agent

  1. Payoff r(s_t,u_t) in some period t is a function of the state vector s_t and control vector u_t
  2. Transition equations are s_{t+1} = g(s_t,u_t)
  3. Assume that u \in U and s \in S
  4. Payoff is bounded: u(s_t,u_t).
24 / 32

Basic theory

Here the current state vector completely summarizes all the information of the past and is all the information the agent needs to make a forward-looking decision

\rightarrow our problem has the Markov property

25 / 32

Basic theory

Here the current state vector completely summarizes all the information of the past and is all the information the agent needs to make a forward-looking decision

\rightarrow our problem has the Markov property

Final two pieces

  1. Stationarity: does not depend explicitly on time
  2. Discounting: \beta \in (0,1), the future matters but not as much as today

Discounting and bounded payoffs ensures total value is bounded

25 / 32

Basic theory

Represent this payoff as \sum_{t=0}^\infty \beta^t r(s_t,u_t)

26 / 32

Basic theory

Represent this payoff as \sum_{t=0}^\infty \beta^t r(s_t,u_t)

The value of the maximized discounted stream of payoffs is \begin{gather} V(s_0) = \max_{u_0 \in U(s_0)} r(s_t,u_t) + \beta \left[\max_{\{u_t\}_{t=1}^\infty} \sum_{t=t}^\infty \beta^t r(s_t,u_t)\right] \notag \\ \text{subject to: } s_{t+1} = g(s_t,u_t) \notag \end{gather}

26 / 32

Basic theory

Represent this payoff as \sum_{t=0}^\infty \beta^t r(s_t,u_t)

The value of the maximized discounted stream of payoffs is \begin{gather} V(s_0) = \max_{u_0 \in U(s_0)} r(s_t,u_t) + \beta \left[\max_{\{u_t\}_{t=1}^\infty} \sum_{t=t}^\infty \beta^t r(s_t,u_t)\right] \notag \\ \text{subject to: } s_{t+1} = g(s_t,u_t) \notag \end{gather}

the terms inside the square brackets is the maximized discounted stream of payoffs beginning at state s_1

26 / 32

Basic theory

This means the problem can be written recursively as \begin{gather} V(s_0) = \max_{u_0 \in U(s_0)} r(s_t,u_t) + \beta\,V(s_1) \\ \text{subject to: } s_{t+1} = g(s_t,u_t) \end{gather}

which is our Bellman (we just exploited Bellman's principle of optimality)

27 / 32

Value function existence and uniqueness

Reformulate the problem as, V(s) = \max_{s' \in \Gamma(s)} r(s,s') + \beta\,V(s'), \,\,\, \forall s \in S where \Gamma(s) is our set of feasible states next period

28 / 32

Value function existence and uniqueness

Reformulate the problem as, V(s) = \max_{s' \in \Gamma(s)} r(s,s') + \beta\,V(s'), \,\,\, \forall s \in S where \Gamma(s) is our set of feasible states next period

There exists a solution to the Bellman under a (particular) set of sufficient conditions:

28 / 32

Value function existence and uniqueness

Reformulate the problem as, V(s) = \max_{s' \in \Gamma(s)} r(s,s') + \beta\,V(s'), \,\,\, \forall s \in S where \Gamma(s) is our set of feasible states next period

There exists a solution to the Bellman under a (particular) set of sufficient conditions:

If the following are true

28 / 32

Value function existence and uniqueness

Reformulate the problem as, V(s) = \max_{s' \in \Gamma(s)} r(s,s') + \beta\,V(s'), \,\,\, \forall s \in S where \Gamma(s) is our set of feasible states next period

There exists a solution to the Bellman under a (particular) set of sufficient conditions:

If the following are true

  1. r(s_t, u_t) is real-valued, continuous and bounded
  2. \beta \in (0,1)
  3. the feasible set of states for next period is non-empty, compact, and continuous
28 / 32

Value function existence and uniqueness

Reformulate the problem as, V(s) = \max_{s' \in \Gamma(s)} r(s,s') + \beta\,V(s'), \,\,\, \forall s \in S where \Gamma(s) is our set of feasible states next period

There exists a solution to the Bellman under a (particular) set of sufficient conditions:

If the following are true

  1. r(s_t, u_t) is real-valued, continuous and bounded
  2. \beta \in (0,1)
  3. the feasible set of states for next period is non-empty, compact, and continuous

then there exists a unique value function V(s) that solves the Bellman equation

28 / 32

Intuitive sketch of the proof

Define an operator T as T(W)(s) = \max_{s' \in \Gamma(s)} r(s,s') + \beta\,W(s'), \,\,\, \forall s \in S

29 / 32

Intuitive sketch of the proof

Define an operator T as T(W)(s) = \max_{s' \in \Gamma(s)} r(s,s') + \beta\,W(s'), \,\,\, \forall s \in S

This operator takes some value function W(s), maximizes it, and returns another T(W)(s)

29 / 32

Intuitive sketch of the proof

Define an operator T as T(W)(s) = \max_{s' \in \Gamma(s)} r(s,s') + \beta\,W(s'), \,\,\, \forall s \in S

This operator takes some value function W(s), maximizes it, and returns another T(W)(s)

It is easy to see that any V(s) that satisfies V(s) = T(V)(s) \,\,\, \forall s \in S solves the Bellman equation

29 / 32

Intuitive sketch of the proof

Define an operator T as T(W)(s) = \max_{s' \in \Gamma(s)} r(s,s') + \beta\,W(s'), \,\,\, \forall s \in S

This operator takes some value function W(s), maximizes it, and returns another T(W)(s)

It is easy to see that any V(s) that satisfies V(s) = T(V)(s) \,\,\, \forall s \in S solves the Bellman equation

Therefore we simply search for the fixed point of T(W) to solve our dynamic problem, but how do we find the fixed point?

29 / 32

Intuitive sketch of the proof

Define an operator T as T(W)(s) = \max_{s' \in \Gamma(s)} r(s,s') + \beta\,W(s'), \,\,\, \forall s \in S

This operator takes some value function W(s), maximizes it, and returns another T(W)(s)

It is easy to see that any V(s) that satisfies V(s) = T(V)(s) \,\,\, \forall s \in S solves the Bellman equation

Therefore we simply search for the fixed point of T(W) to solve our dynamic problem, but how do we find the fixed point?

First we must show that a way exists by showing that T(W) is a contraction:
as we iterate using the T operator, we will get closer and closer to the fixed point

29 / 32

Intuitive sketch of the proof

Blackwell's sufficient conditions for a contraction are

30 / 32

Intuitive sketch of the proof

Blackwell's sufficient conditions for a contraction are

  1. Monotonicity: if W(s) \geq Q(s) \,\,\, \forall s \in S, then T(W)(s) \geq T(Q)(s) \,\,\, \forall s \in S
30 / 32

Intuitive sketch of the proof

Blackwell's sufficient conditions for a contraction are

  1. Monotonicity: if W(s) \geq Q(s) \,\,\, \forall s \in S, then T(W)(s) \geq T(Q)(s) \,\,\, \forall s \in S
  2. Discounting: there exists a \beta \in (0,1) such that T(W+k)(s) \leq T(W)(s) + \beta\,k
30 / 32

Intuitive sketch of the proof

Blackwell's sufficient conditions for a contraction are

  1. Monotonicity: if W(s) \geq Q(s) \,\,\, \forall s \in S, then T(W)(s) \geq T(Q)(s) \,\,\, \forall s \in S
  2. Discounting: there exists a \beta \in (0,1) such that T(W+k)(s) \leq T(W)(s) + \beta\,k

Monotonicity holds under our maximization

30 / 32

Intuitive sketch of the proof

Blackwell's sufficient conditions for a contraction are

  1. Monotonicity: if W(s) \geq Q(s) \,\,\, \forall s \in S, then T(W)(s) \geq T(Q)(s) \,\,\, \forall s \in S
  2. Discounting: there exists a \beta \in (0,1) such that T(W+k)(s) \leq T(W)(s) + \beta\,k

Monotonicity holds under our maximization

Discounting reflects that we must be discounting the future

30 / 32

Intuitive sketch of the proof

Blackwell's sufficient conditions for a contraction are

  1. Monotonicity: if W(s) \geq Q(s) \,\,\, \forall s \in S, then T(W)(s) \geq T(Q)(s) \,\,\, \forall s \in S
  2. Discounting: there exists a \beta \in (0,1) such that T(W+k)(s) \leq T(W)(s) + \beta\,k

Monotonicity holds under our maximization

Discounting reflects that we must be discounting the future

If these two conditions hold then we have a contraction with modulus \beta

30 / 32

Intuitive sketch of the proof

Blackwell's sufficient conditions for a contraction are

  1. Monotonicity: if W(s) \geq Q(s) \,\,\, \forall s \in S, then T(W)(s) \geq T(Q)(s) \,\,\, \forall s \in S
  2. Discounting: there exists a \beta \in (0,1) such that T(W+k)(s) \leq T(W)(s) + \beta\,k

Monotonicity holds under our maximization

Discounting reflects that we must be discounting the future

If these two conditions hold then we have a contraction with modulus \beta

Why do we care this is a contraction?

30 / 32

Intuitive sketch of the proof

So we can take advantage of the contraction mapping theorem which states:

31 / 32

Intuitive sketch of the proof

So we can take advantage of the contraction mapping theorem which states:

  1. T has a unique fixed point
  2. T(V^*) = V^*
  3. We can start from any arbitrary initial function W, iterate using T and reach the fixed point
31 / 32

Next up

Next: numerical methods for discrete time dynamics

32 / 32

Roadmap

  1. Review Markov chains and dynamic models
  2. Review theory for numerical methods
2 / 32
Paused

Help

Keyboard shortcuts

, , Pg Up, k Go to previous slide
, , Pg Dn, Space, j Go to next slide
Home Go to first slide
End Go to last slide
Number + Return Go to specific slide
b / m / f Toggle blackout / mirrored / fullscreen mode
c Clone slideshow
p Toggle presenter mode
t Restart the presentation timer
?, h Toggle this help
Esc Back to slideshow