For our purposes, let’s define it as:
consistent and systematic deviation from the expected norm stemming from a flaw in the design of study.
“The order of presenting the treatments affects the dependent variable.” (Cozby, 137)
(sometimes called sequence effects)
People treat the most recently presented things differently from the others.
People treat the things presented first differently than they would the others.
Sometimes simple randomization won’t work. Maybe there are repeated measures and you don’t want them to be presented back-to-back, for example. Or maybe the order is an effect that you’re actually interested in.
A latin square design is sometimes a good way to go.
First | Second | Third | Fourth |
---|---|---|---|
A | B | C | D |
B | C | D | A |
C | D | A | B |
This is probably the most commonly asked question. First we should cover a few preliminaries:
From Pollard and Richardson (1987):
“Whenever you find a null result and it is interesting to you that the result is null, you should always indicate the sensitivity of your analysis.” (Dienes, p.68)
Power
Confidence Intervals
Finding an effect significantly different from another reference point.
-Neyman and Pearson would argue that you need to adequately define the conditions under which you will stop collecting data a priori.
Type I Error: Wrongly claiming something to be true, useful or knowable.
Type II Error: Wrongly claiming something to be false, useless or unknowable.
\(\alpha\) = The part of a distribution that is so extreme that we can reject the likelihood of it happening (usually in alignment with our maximum accepted p-value.)
\(\beta\) = When the null is false, \(\beta\) is the proportion of times that we accept it, even though it’s false.
types of errors
(from Dienes, p.63)
Decision | \(\ H_0\) true | \(\ H_0\) false |
---|---|---|
Accept \(\ H_0\) | 3800 | 500 |
Reject \(\ H_0\) | 200 | 500 |
Totals | 4000 | 1000 |
This means we consider only the cases when we have rejected the null. (200/700, or 29%)
The point: Using only a significance level of 5% does not guarantee that only 5% of all published significant results are an error.
Estimate the size of effect that you think is interesting, given that your theory is true.
Estimate the amount of noise that your data will have.
If you plan it ahead of time, then you can stick with your original \(\alpha\).
if you would like to test other things, as you go, you need to correct for multiple tests.
(taken from Dienes, p.76)