Dr Adequate
Banned
- Joined
- Aug 31, 2004
- Messages
- 17,766
I never did statistics past A Level, which has worked fine so far. However, I now find myself scratching my little head.
Suppose I have a hypothesis that a coin is fair. I toss it a thousand times, and 990 times it comes down heads. I can work out the probabilty of that, and say: now, this would only happen one time in (some large number I can't be bothered to calculate) if the hypothesis is correct, therefore the hypothesis only stands those odds of being correct.
Now, this seems to me to be fair and reasonable.
Only I can also look at any sequence of a thousand tosses, and say: now, this would only happen one time in 21000 if the hypothesis is correct, therefore the hypothesis only stands those odds of being correct.
Now, the reasoning is the same in both cases, but in the second case it obviously doesn't work. Therefore, something is wrong with the reasoning.
I can see why the reasoning must be untrue in the second case: for the statement "this would only happen one time in 21000 if the hypothesis is correct" would be true however the coins fell, whether or not the coin was fair, and so cannot be held to contradict the hypothesis.
Hence the reasoning in the first case is incorrect and incomplete. How do I remedy this? Could I see the reasoning and the math?
There must be something somewhere which explains how one's definition of "improbable" relates to the testing of a statistical hypothesis: a formal and well-founded way to tell how significant to a statistical hypothesis an improbable event really is. All events are improbable to some degree, depending on how you do the counting. What is the right way to do the counting, and how is this demonstrated?
Thanks.
Suppose I have a hypothesis that a coin is fair. I toss it a thousand times, and 990 times it comes down heads. I can work out the probabilty of that, and say: now, this would only happen one time in (some large number I can't be bothered to calculate) if the hypothesis is correct, therefore the hypothesis only stands those odds of being correct.
Now, this seems to me to be fair and reasonable.
Only I can also look at any sequence of a thousand tosses, and say: now, this would only happen one time in 21000 if the hypothesis is correct, therefore the hypothesis only stands those odds of being correct.
Now, the reasoning is the same in both cases, but in the second case it obviously doesn't work. Therefore, something is wrong with the reasoning.
I can see why the reasoning must be untrue in the second case: for the statement "this would only happen one time in 21000 if the hypothesis is correct" would be true however the coins fell, whether or not the coin was fair, and so cannot be held to contradict the hypothesis.
Hence the reasoning in the first case is incorrect and incomplete. How do I remedy this? Could I see the reasoning and the math?
There must be something somewhere which explains how one's definition of "improbable" relates to the testing of a statistical hypothesis: a formal and well-founded way to tell how significant to a statistical hypothesis an improbable event really is. All events are improbable to some degree, depending on how you do the counting. What is the right way to do the counting, and how is this demonstrated?
Thanks.