Skip to content

Commit 85183d4

Browse files
Tom's July 1 edits of two Wald lectures
1 parent 9e26a71 commit 85183d4

2 files changed

Lines changed: 20 additions & 6 deletions

File tree

lectures/wald_friedman.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -276,7 +276,7 @@ Here is how Wald introduces the notion of a sequential test
276276
277277
## Wald's Sequential Formulation
278278

279-
In contradistinction to Neyman and Pearson's formulation of the problemm, in Wald's formulation
279+
In contradistinction to Neyman and Pearson's formulation of the problem, in Wald's formulation
280280

281281

282282
- The sample size $n$ is not fixed but rather a random variable.

lectures/wald_friedman_2.md

Lines changed: 19 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -43,9 +43,23 @@ In {doc}`this lecture <wald_friedman>`, we described how Abraham Wald {cite}`W
4343
Wald's idea of formulating the problem sequentially created links to the **dynamic programming** that Richard Bellman developed in the 1950s.
4444
```
4545

46-
In this lecture, we reformulate the problem in a way that invites us to apply dynamic programming.
46+
As we learned in {doc}`probability with matrices <prob_matrix>` and {doc}`two meanings of probability<prob_meaning>`, a frequentist statistician views a probability distribution as measuring relative frequencies of a statistic that he anticipates constructing from a very long sequence of i.i.d. draws from a known probability distribution.
4747

48-
We'll do this by transforming our point of view from the 'objective' frequentist perspective of {doc}`this lecture <wald_friedman>` to the explicitly 'subjective' perspective taken by a Bayesian decision maker who is equipped with
48+
That known probability distribution is his 'hypothesis'.
49+
50+
A frequentist statistician studies the distribution of that statistic under that known probability distribution
51+
52+
* when the distribution is a member of a set of parameterized probability distribution, his hypothesis takes the form of a particular parameter vector.
53+
* this is what we mean when we say that the frequentist statistician 'conditions on the parameters'
54+
* he regards the parameters that are fixed numbers, known to nature, but not to him.
55+
* the statistician copes with his ignorane of those parameters by constructing the type I and type II errors associated with frequentist hypothesis testing.
56+
57+
In this lecture, we reformulate Friedman and Wald's problem by transforming our point of view from the 'objective' frequentist perspective of {doc}`this lecture <wald_friedman>` to an explicitly 'subjective' perspective taken by a Bayesian decision maker who regards parameters not as fixed numbers but as (hidden) random variables that are jointly distributed with the random variables that can be observed by sampling from that joint distribution.
58+
59+
To form that joint distribution, the Bayesian statistician supplements the conditional distributions used by the frequentist statistician with
60+
a prior probability distribution over the parameters that representive his personal, subjective opinion about those them.
61+
62+
To proceed in the way, we endow our decision maker with
4963

5064
- an initial prior subjective probability $\pi_{-1} \in (0,1)$ that nature uses to generate $\{z_k\}$ as a sequence of i.i.d. draws from $f_1$ rather than $f_0$.
5165
- faith in Bayes' law as a way to revise his subjective beliefs as observations on $\{z_k\}$ sequence arrive each period.
@@ -62,10 +76,10 @@ In our {doc}`previous frequentist version <wald_friedman>`, key ideas in play we
6276
- A **uniformly most powerful test**
6377

6478
In this lecture about a Bayesian reformulation of the problem, additional ideas at work are
65-
- an initial prior probability $\pi_{-1}$ that
66-
- a sequence of posterior probabilities
79+
- an initial prior probability $\pi_{-1}$ that model $f_1$ generates the data
6780
- Bayes' Law
68-
- Dynamic programming
81+
- a sequence of posterior probabilities that model $f_1$ is generating the data
82+
- dynamic programming
6983

7084

7185
This lecture uses ideas studied in {doc}`this lecture <likelihood_ratio_process>`, {doc}`this lecture <likelihood_bayes>`, and {doc}`this lecture <exchangeable>`.

0 commit comments

Comments
 (0)