## Still not sure what you are aiming at… [General Statistics]

Hi ElMaestro,

» 1. Let us look at the wikipedia page for the t test:

» "Most test statistics have the form t = Z/s, where Z and s are functions of the data."

OK, so far.

» 2. For the t-distribution, here Z=sample mean - mean and s=sd/sqrt(n)

Wait a minute. You are referring to the one-sample

» 3. Why are Z and s independent in this case?

Here we know the population mean. Hence, the numerator depends on the sample mean and the denominator on the sample’s standard error. They are independent indeed.

I added another plot to the code of this post.

A modified plot of 5,000 samples to the right.

» Or more generally, and for me much more importantly, if we have two functions (f and g, or Z and s), then which properties of such functions or their input would render them independent??

» Wikipedia links to a page about independence, key here is: […]

Yep.

» I am fully aware that when we simulate a normal dist. with some mean and some variance, then that defines their expected estimates in a sample. I.e. if a sample has a mean that is higher than the simulated mean, then that does not necessarily mean the sampled sd is higher (or lower, for that matter, that was where I was going with "perturbation"). It sounds right to think of the two as independent, in that case.

Correct. Anything is possible.

» Now, how about the general case, for example if we know nothing about the nature of the sample, but just look at any two functions of the sample? What property would we look for in those two functions to think they are independent?

» A general understanding of the idea of independence of any two quantities derived from a sample, that is what I am looking for; point #3 above defines my question.

Still not sure whether I understand you~~ correctly ~~ at all. Think about the general formulation of a test statistic from above $$t=\frac{Z}{s},$$where \(Z\) and \(s\) are functions of the data.

I think that this formulation is unfortunate because it has neither to do anything with the standard normal distribution \(Z\) nor the sample standard deviation \(s\). For continuous variables I would prefer sumfink like$$test\;statistic=\frac{measure\;of\;location}{measure\;of\;dispersion}$$for clarity. If a test would be constructed in such a way that the independence is not correctly represented it would be a piece of shit.

» 1. Let us look at the wikipedia page for the t test:

» "Most test statistics have the form t = Z/s, where Z and s are functions of the data."

OK, so far.

» 2. For the t-distribution, here Z=sample mean - mean and s=sd/sqrt(n)

Wait a minute. You are referring to the one-sample

*t*-test, right? At the Assumptions we find$$t=\frac{Z}{s}=\frac{\bar{X}-\mu}{\hat{\sigma}/\sqrt{n}}$$That’s a little bit strange because WP continues with\(\hat{\sigma}\) is the estimate of the standard deviation of the population

I beg your pardon? Most of my textbooks give the same formula but with \(s\) in the denominator as the sample standard deviation. Of course, \(s/\sqrt{n}\) is the standard error and sometimes we find \(t=\frac{\bar{X}-\mu}{\textrm{SE}}\) instead. Nevertheless, in further down we find$$t=\frac{\bar{x}-\mu_0}{s/\sqrt{n}}$$THX a lot, soothing!» 3. Why are Z and s independent in this case?

Here we know the population mean. Hence, the numerator depends on the sample mean and the denominator on the sample’s standard error. They are independent indeed.

I added another plot to the code of this post.

A modified plot of 5,000 samples to the right.

» Or more generally, and for me much more importantly, if we have two functions (f and g, or Z and s), then which properties of such functions or their input would render them independent??

» Wikipedia links to a page about independence, key here is: […]

Yep.

» I am fully aware that when we simulate a normal dist. with some mean and some variance, then that defines their expected estimates in a sample. I.e. if a sample has a mean that is higher than the simulated mean, then that does not necessarily mean the sampled sd is higher (or lower, for that matter, that was where I was going with "perturbation"). It sounds right to think of the two as independent, in that case.

Correct. Anything is possible.

» Now, how about the general case, for example if we know nothing about the nature of the sample, but just look at any two functions of the sample? What property would we look for in those two functions to think they are independent?

» A general understanding of the idea of independence of any two quantities derived from a sample, that is what I am looking for; point #3 above defines my question.

Still not sure whether I understand you

I think that this formulation is unfortunate because it has neither to do anything with the standard normal distribution \(Z\) nor the sample standard deviation \(s\). For continuous variables I would prefer sumfink like$$test\;statistic=\frac{measure\;of\;location}{measure\;of\;dispersion}$$for clarity. If a test would be constructed in such a way that the independence is not correctly represented it would be a piece of shit.

—

Helmut Schütz

The quality of responses received is directly proportional to the quality of the question asked. 🚮

Science Quotes

*Dif-tor heh smusma*🖖_{}Helmut Schütz

The quality of responses received is directly proportional to the quality of the question asked. 🚮

Science Quotes

### Complete thread:

- Statistical independence, what is it? I mean really, what is it?? ElMaestro 2020-06-27 21:35 [General Statistics]
- Die don’t remember the last roll. Really. Helmut 2020-06-28 13:35
- Die don’t remember the last roll. Really. ElMaestro 2020-06-28 14:45
- Die don’t remember the last roll. Really. Helmut 2020-06-28 15:36
- Still none the wiser ElMaestro 2020-06-28 18:20
- You’ve lost me now. Helmut 2020-06-28 21:55
- Worded differently ElMaestro 2020-06-29 08:30
- Still not sure what you are aiming at…Helmut 2020-06-29 16:46
- Still not sure what you are aiming at… ElMaestro 2020-06-30 00:55
- Confuse-a-Cat Helmut 2020-06-30 11:33
- Confuse-a-Cat ElMaestro 2020-06-30 13:07
- Confuse-a-Cat Helmut 2020-06-30 14:27
- pseudorandom and linear independence mittyri 2020-07-01 00:04

- Confuse-a-Cat ElMaestro 2020-06-30 13:07

- Confuse-a-Cat Helmut 2020-06-30 11:33

- Still not sure what you are aiming at… ElMaestro 2020-06-30 00:55

- Still not sure what you are aiming at…Helmut 2020-06-29 16:46

- Worded differently ElMaestro 2020-06-29 08:30

- You’ve lost me now. Helmut 2020-06-28 21:55

- Still none the wiser ElMaestro 2020-06-28 18:20

- Die don’t remember the last roll. Really. Helmut 2020-06-28 15:36

- Die don’t remember the last roll. Really. ElMaestro 2020-06-28 14:45
- Statistical independence, what is it? I mean really, what is it?? martin 2020-07-01 08:40
- Statistical independence, what is it? I mean really, what is it?? ElMaestro 2020-07-01 09:42
- Statistical independence, what is it? I mean really, what is it?? martin 2020-07-01 10:07

- Statistical independence, what is it? I mean really, what is it?? ElMaestro 2020-07-01 09:42

- Die don’t remember the last roll. Really. Helmut 2020-06-28 13:35