Closed mine-cetinkaya-rundel closed 5 years ago
If this change is implemented it will also affect the following line,
where we could do away with the approximation since by definition SE is an approximation, and say
SE = \sqrt{\frac{\wsjebolapollprop{} (1 - \wsjebolapollprop{})}{\wsjebolapollsize{}}}
Let's catch up on this in our next meeting; I've added a note there so this doesn't get lost.
https://github.com/OpenIntroOrg/openintro-statistics/blob/b7e661929a25ca0f7fb530139e1c5aa015f1f459/ch_foundations_for_inf/TeX/ch_foundations_for_inf.tex#L212-L218
A more accurate definition would be that the variability of the sample statistic is called the standard error when calculated using sample statistics, since it's called the standard deviation if calculated using true population parameters.
Not sure if going into that distinction is worthwhile here, since the latter case rarely happens, but the definition here makes it sound like the variability of a sample statistic is always called standard error, which is not accurate.
Relatedly, the formula below is not correct:
https://github.com/OpenIntroOrg/openintro-statistics/blob/b7e661929a25ca0f7fb530139e1c5aa015f1f459/ch_foundations_for_inf/TeX/ch_foundations_for_inf.tex#L284
If using p, this should be called sigma. If using p-hat, it would be called SE.
I would suggest the following changes to address these two concerns:
We typically measure the variability of a sampling distribution, or the variability of a point estimate, with \termsub{standard error}{standard error (SE)}, and the notation $SE_{\hat{p}}$ is used for the standard error associated with the sample proportion.
sigma_{\hat{p}} &= \sqrt{\frac{p (1 - p)}{n}}
Often we don't know p, so we can't directly calculate sigma_{\hat{p}}, and instead estimate it with SE_{\hat{p}} &= \sqrt{\frac{\hat{p} (1 - \hat{p})}{n}}.