You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- In the first figure, $\log L(w^t)$ diverges to $\infty$ because $K_g > K_f$.
752
761
- In the second figure, we still have $K_g > K_f$, but the difference is smaller, so $L(w^t)$ diverges to infinity at a slower pace.
753
762
- In the last figure, $\log L(w^t)$ diverges to $-\infty$ because $K_g < K_f$.
754
-
- The black dotted line, $t \cdot \left(KL(h,g) - KL(h, f)\right)$, closely fits the paths verifying {eq}`eq:kl_likelihood_link`.
763
+
- The black dotted line, $t \left(KL(h,g) - KL(h, f)\right)$, closely fits the paths verifying {eq}`eq:kl_likelihood_link`.
755
764
756
765
These observations align with the theory.
757
766
@@ -781,8 +790,7 @@ We assume that $f$ and $g$ both put positive probabilities on the same intervals
781
790
782
791
783
792
784
-
In the simulations below, we specify that $f$ is a $\text{Beta}(1, 1)$ distribution and that $g$ is $\text{Beta}(3, 1.2)$ distribution,
785
-
just as we did often earlier in this lecture.
793
+
In the simulations below, we specify that $f$ is a $\text{Beta}(1, 1)$ distribution and that $g$ is $\text{Beta}(3, 1.2)$ distribution.
786
794
787
795
We consider two alternative timing protocols.
788
796
@@ -1028,7 +1036,7 @@ plt.tight_layout()
1028
1036
plt.show()
1029
1037
```
1030
1038
1031
-
To the left of the green vertical line $f < g $, so $l_t < 1$; therefore a $w_t$ that falls to the left of the green line is classified as a type $g$ individual.
1039
+
To the left of the green vertical line $g < f$, so $l_t < 1$; therefore a $w_t$ that falls to the left of the green line is classified as a type $g$ individual.
1032
1040
1033
1041
* The shaded orange area equals $\beta$ -- the probability of classifying someone as a type $g$ individual when it is really a type $f$ individual.
1034
1042
@@ -1226,7 +1234,7 @@ Because in general $KL(f, g) \neq KL(g, f)$, KL divergence is not symmetric, but
1226
1234
As {eq}`eq:js_divergence` shows, the Jensen-Shannon divergence computes average of the KL divergence of $f$ and $g$ with respect to a particular reference distribution $m$ defined below the equation.
1227
1235
```
1228
1236
1229
-
Now let's create a comparison table showing KL divergence, Jensen-Shannon divergence, and Chernoff entropy
1237
+
Now let's create a comparison table showing KL divergence, Jensen-Shannon divergence, and Chernoff entropy for a set of pairs of Beta distributions.
We find that $KL(f,g) > KL(g,f)$ and $KL(g,h) > KL(f,h)$.
1984
+
We find that $KL(f,g) > KL(g,f)$ and $KL(h,g) > KL(h,f)$.
1977
1985
1978
1986
The first inequality tells us that the average "surprise" or "inefficiency" of using belief $g$ when nature chooses $f$ is greater than the "surprise" of using belief $f$ when nature chooses $g$.
We find that in the first case, $KL(f,g) \approx KL(g,f)$ and both are relatively small, so although either agent 1 or agent 2 will eventually consume everything, convergence displaying in first two panels on the top is pretty slowly.
2112
2120
2113
-
In the first two panels at the bottom, we see convergence occurring faster because the divergence gap $KL(f, g) > KL(g, f)$ is larger (as indicated by the black dashed line).
2121
+
In the first two panels at the bottom, we see convergence occurring faster (as indicated by the black dashed line) because the divergence gaps $KL(f, g)$ and $KL(g, f)$ are larger.
2114
2122
2115
-
We see faster convergence in the first panel at the bottom when nature chooses $f$ than in the second panel where nature chooses $g$.
2123
+
Since $KL(f,g) > KL(g,f)$, we see faster convergence in the first panel at the bottom when nature chooses $f$ than in the second panel where nature chooses $g$.
2116
2124
2117
2125
This ties in nicely with {eq}`eq:kl_likelihood_link`.
0 commit comments