You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: lectures/wald_friedman.md
+17-14Lines changed: 17 additions & 14 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -124,7 +124,7 @@ We'll formulate the problem using dynamic programming.
124
124
The following presentation of the problem closely follows Dmitri
125
125
Berskekas's treatment in **Dynamic Programming and Stochastic Control** {cite}`Bertekas75`.
126
126
127
-
A decision-maker observes a sequence of draws of a random variable $z$.
127
+
A decision-maker can observe a sequence of draws of a random variable $z$.
128
128
129
129
He (or she) wants to know which of two probability distributions $f_0$ or $f_1$ governs $z$.
130
130
@@ -137,19 +137,23 @@ random variables is also independently and identically distributed (IID).
137
137
But the observer does not know which of the two distributions generated the sequence.
138
138
139
139
For reasons explained in [Exchangeability and Bayesian Updating](https://python.quantecon.org/exchangeable.html), this means that the sequence is not
140
-
IID and that the observer has something to learn, even though he knows both $f_0$ and $f_1$.
140
+
IID.
141
141
142
-
The decision maker chooses a number of draws (i.e., random samples from the unknown distribution) and uses them to decide
142
+
The observer has something to learn, namely, whether the observations are drawn from $f_0$ or from $f_1$.
143
+
144
+
The decision maker wants to decide
143
145
which of the two distributions is generating outcomes.
144
146
145
-
He starts with prior
147
+
We adopt a Bayesian formulation.
148
+
149
+
The decision maker begins with a prior probability
146
150
147
151
$$
148
152
\pi_{-1} =
149
153
\mathbb P \{ f = f_0 \mid \textrm{ no observations} \} \in (0, 1)
150
154
$$
151
155
152
-
After observing $k+1$ observations $z_k, z_{k-1}, \ldots, z_0$, he updates this value to
156
+
After observing $k+1$ observations $z_k, z_{k-1}, \ldots, z_0$, he updates his personal probability that the observations are described by distribution $f_0$ to
153
157
154
158
$$
155
159
\pi_k = \mathbb P \{ f = f_0 \mid z_k, z_{k-1}, \ldots, z_0 \}
@@ -252,15 +256,15 @@ So when we treat $f=f_0$ as the null hypothesis
252
256
253
257
### Intuition
254
258
255
-
Let's try to guess what an optimal decision rule might look like before we go further.
259
+
Before proceeding, let's try to guess what an optimal decision rule might look like.
256
260
257
261
Suppose at some given point in time that $\pi$ is close to 1.
258
262
259
263
Then our prior beliefs and the evidence so far point strongly to $f = f_0$.
260
264
261
265
If, on the other hand, $\pi$ is close to 0, then $f = f_1$ is strongly favored.
262
266
263
-
Finally, if $\pi$ is in the middle of the interval $[0, 1]$, then we have little information in either direction.
267
+
Finally, if $\pi$ is in the middle of the interval $[0, 1]$, then we are confronted with more uncertainty.
264
268
265
269
This reasoning suggests a decision rule such as the one shown in the figure
266
270
@@ -270,8 +274,7 @@ This reasoning suggests a decision rule such as the one shown in the figure
270
274
271
275
As we'll see, this is indeed the correct form of the decision rule.
272
276
273
-
The key problem is to determine the threshold values $\alpha, \beta$,
274
-
which will depend on the parameters listed above.
277
+
Our problem is to determine threshold values $\alpha, \beta$ that somehow depend on the parameters described above.
275
278
276
279
You might like to pause at this point and try to predict the impact of a
277
280
parameter such as $c$ or $L_0$ on $\alpha$ or $\beta$.
@@ -326,7 +329,7 @@ where $\pi \in [0,1]$ and
326
329
$f_0$ (i.e., the cost of making a type II error).
327
330
- $\pi L_1$ is the expected loss associated with accepting
328
331
$f_1$ (i.e., the cost of making a type I error).
329
-
- $h(\pi) := c + \mathbb E [J(\pi')]$ the continuation value; i.e.,
332
+
- $h(\pi) := c + \mathbb E [J(\pi')]$; this is the continuation value; i.e.,
330
333
the expected cost associated with drawing one more $z$.
331
334
332
335
The optimal decision rule is characterized by two numbers $\alpha, \beta \in (0,1) \times (0,1)$ that satisfy
@@ -354,7 +357,7 @@ $$
354
357
Our aim is to compute the value function $J$, and from it the associated cutoffs $\alpha$
355
358
and $\beta$.
356
359
357
-
To make our computations simpler, using {eq}`optdec`, we can write the continuation value $h(\pi)$ as
360
+
To make our computations manageable, using {eq}`optdec`, we can write the continuation value $h(\pi)$ as
0 commit comments