Skip to content
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 5 additions & 5 deletions lectures/lln_clt.md
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ print(X)
```

In this setting, the LLN tells us if we flip the coin many times, the fraction
of heads that we see will be close to the mean $p$.
of heads that we see will be close to the mean $p$. We use $n$ to represent the number of times the coin is flipped.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@SylviaZhaooo Would you mind to put the second sentence on a separate line -- this is our (unusual) convention for the lectures.

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sure!


Let's check this:

Expand Down Expand Up @@ -286,7 +286,7 @@ as expected.

Let's vary `n` to see how the distribution of the sample mean changes.

We will use a violin plot to show the different distributions.
We will use a [violin plot](https://intro.quantecon.org/prob_dist.html#violin-plots) to show the different distributions.

Each distribution in the violin plot represents the distribution of $X_n$ for some $n$, calculated by simulation.

Expand Down Expand Up @@ -357,7 +357,7 @@ This means that the distribution of $\bar X_n$ does not eventually concentrate o

Hence the LLN does not hold.

The LLN fails to hold here because the assumption $\mathbb E|X| = \infty$ is violated by the Cauchy distribution.
The LLN fails to hold here because the assumption $\mathbb E|X| < \infty$ is violated by the Cauchy distribution.

+++

Expand Down Expand Up @@ -438,7 +438,7 @@ Here $\stackrel { d } {\to} N(0, \sigma^2)$ indicates [convergence in distributi

The striking implication of the CLT is that for **any** distribution with
finite [second moment](https://en.wikipedia.org/wiki/Moment_(mathematics)), the simple operation of adding independent
copies **always** leads to a Gaussian curve.
copies **always** leads to a Gaussian(Normal) curve.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
copies **always** leads to a Gaussian(Normal) curve.
copies **always** leads to a Gaussian (Normal) curve.

This is a really minor point but @jstac re: style I think there should be space between Gaussian and (Normal). Do you agree?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes please, thanks @mmcky and @SylviaZhaooo




Expand Down Expand Up @@ -503,7 +503,7 @@ The fit to the normal density is already tight and can be further improved by in
```{exercise}
:label: lln_ex1

Repeat the simulation [above1](sim_one) with the [Beta distribution](https://en.wikipedia.org/wiki/Beta_distribution).
Repeat the simulation [above](sim_one) with the [Beta distribution](https://en.wikipedia.org/wiki/Beta_distribution).

You can choose any $\alpha > 0$ and $\beta > 0$.
```
Expand Down