You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: lectures/mccall_model.md
+13-13Lines changed: 13 additions & 13 deletions
Original file line number
Diff line number
Diff line change
@@ -91,7 +91,7 @@ economists to inject randomness into their models.)
91
91
92
92
In this lecture, we adopt the following simple environment:
93
93
94
-
* $\{s_t\}$ is IID, with $q(s)$ being the probability of observing state $s$ in $\mathbb{S}$ at each point in time,
94
+
* $\{s_t\}$ is IID, with $q(s)$ being the probability of observing state $s$ in $\mathbb{S}$ at each point in time,
95
95
* the agent observes $s_t$ at the start of $t$ and hence knows
96
96
$w_t = w(s_t)$,
97
97
* the set $\mathbb S$ is finite.
@@ -120,7 +120,7 @@ The variable $y_t$ is income, equal to
120
120
* unemployment compensation $c$ when unemployed
121
121
122
122
The worker knows that $\{s_t\}$ is IID with common
123
-
distribution $q$ and uses knowledge when he or she computes mathematical expectations of various random variables that are functions of
123
+
distribution $q$ and uses knowledge when he or she computes mathematical expectations of various random variables that are functions of
124
124
$s_t$.
125
125
126
126
### A Trade-Off
@@ -134,7 +134,7 @@ To decide optimally in the face of this trade-off, we use dynamic programming.
134
134
135
135
Dynamic programming can be thought of as a two-step procedure that
136
136
137
-
1. first assigns values to "states"
137
+
1. first assigns values to "states"
138
138
1. then deduces optimal actions given those values
139
139
140
140
We'll go through these steps in turn.
@@ -160,16 +160,16 @@ Let $v^*(s)$ be the optimal value of the problem when $s \in \mathbb{S}$ for a
160
160
161
161
162
162
163
-
Thus, the function $v^*(s)$ is the maximum value of objective
163
+
Thus, the function $v^*(s)$ is the maximum value of objective
164
164
{eq}`objective` for a previously unemployed worker who has offer $w(s)$ in hand and has yet to choose whether to accept it.
165
165
166
166
Notice that $v^*(s)$ is part of the **solution** of the problem, so it isn't obvious that it is a good idea to start working on the problem by focusing on $v^*(s)$.
167
167
168
168
There is a chicken and egg problem: we don't know how to compute $v^*(s)$ because we don't yet know
169
169
what decisions are optimal and what aren't!
170
170
171
-
But it turns out to be a really good idea by asking what properties the optimal value function $v^*(s)$ must have in order it
172
-
to qualify as an optimal value function.
171
+
But it turns out to be a really good idea by asking what properties the optimal value function $v^*(s)$ must have in order it
172
+
to qualify as an optimal value function.
173
173
174
174
Think of $v^*$ as a function that assigns to each possible state
175
175
$s$ the maximal expected discounted income stream that can be obtained with that offer in
@@ -192,7 +192,7 @@ for every possible $s$ in $\mathbb S$.
192
192
Notice how the function $v^*(s)$ appears on both the right and left sides of equation {eq}`odu_pv` -- that is why it is called
193
193
a **functional equation**, i.e., an equation that restricts a **function**.
194
194
195
-
This important equation is a version of a **Bellman equation**, an equation that is
195
+
This important equation is a version of a **Bellman equation**, an equation that is
196
196
ubiquitous in economic dynamics and other fields involving planning over time.
197
197
198
198
The intuition behind it is as follows:
@@ -218,7 +218,7 @@ Once we have this function in hand we can figure out how behave optimally (i.e.
218
218
219
219
All we have to do is select the maximal choice on the r.h.s. of {eq}`odu_pv`.
220
220
221
-
The optimal action in state $s$ can be thought of as a part of a **policy** that maps a
221
+
The optimal action in state $s$ can be thought of as a part of a **policy** that maps a
222
222
state into an action.
223
223
224
224
Given *any* $s$, we can read off the corresponding best choice (accept or
@@ -351,7 +351,7 @@ Moreover, it's immediate from the definition of $T$ that this fixed
351
351
point is $v^*$.
352
352
353
353
A second implication of the Banach contraction mapping theorem is that
354
-
$\{ T^k v \}$ converges to the fixed point $v^*$ regardless of the initial
354
+
$\{ T^k v \}$ converges to the fixed point $v^*$ regardless of the initial
355
355
$v \in \mathbb R^n$.
356
356
357
357
### Implementation
@@ -386,7 +386,7 @@ We are going to use Numba to accelerate our code.
386
386
387
387
* See, in particular, the discussion of `@jitclass` in [our lecture on Numba](https://python-programming.quantecon.org/numba.html).
388
388
389
-
The following helps Numba by providing some information about types
389
+
The following helps Numba by providing some information about types
0 commit comments