Академический Документы
Профессиональный Документы
Культура Документы
Chapter 11
measurement error)
NSHT allows us to make a decision:
Is the difference large enough to be highly unlikely
(e.g., p < .05) to be due to chance? OR, is the difference small enough that it could reflect error variance?
error variance even if the independent variable has no effect (H0). If the obtained difference exceeds that amount, reject H0. (the difference is highly unlikely [e.g., p <
.05] to be due to chance)
Otherwise, fail to reject H0 (no significant
difference).
Any differences may have been due to chance (p > .05) Still, we cannot logically conclude that there is no
true Concluding that there is an effect when there actually is not Alpha level (a) is the probability of making a Type I error.
Type II Errors:
Fail to reject H0 (the null hypothesis) when it is
actually false Concluding that there is not an effect when there actually is Beta () is the probability of making a Type II error
Type II error
Statistical Power
The power of a test is the probability that the test will
Effect size Sample size Alpha level Type of test (one- vs. two-tailed)
Influences on Power
Power is increased with larger effect sizes larger sample sizes larger (less stringent) alpha levels (e.g., a = .05) one-tailed (instead of two-tailed tests) tests
Power Illustration
Sampling distribution if H0 is true Sampling distribution if H0 is false
Power = 1
H1
H1
Power (1-b)
H1
H1
Power (1-b)
Effect Size
Statistical significance (rejecting H0) does not
necessarily tell you about the size or practical significance of an effect Statistically significant effects are not always practically significant!
Effect size measures the absolute magnitude of a
proportion of variance in the dependent variable that is due to the independent variable; ranges from .00 to 1.00
Mean difference effect size (Cohens d) size of
the difference between two means relative to the standard deviation of the data; typically ranges from -3.0 to +3.0
d = .20
d = .50 d = .80
Small effect
Medium effect Large effect
Power examples
Large, |r|=.50
Small, d =.20 t-test with equal Ns Medium, d =.50 Large, d =.80 Small, f = .10 2x2 ANOVA Medium, f = .25 (interaction term) Large, f = .40
N = 26
N = 788 N = 128 N = 52 N = 787 N = 128 N = 52
(small) effect In reality, most null hypotheses may be false (Loftus, 1996)
Example H0: m1 = m2 = m3 = m4 This means there are ZERO differences (not even .00001)
in the population
significance. Use meta-analysis of effect sizes across studies, when possible Try to maintain adequate power (ideally > .80). Report plots and graphs of data, when appropriate. Report confidence intervals (CIs)
E.g., We can be 95% confident that the population
mean difference lies between 0.04 and 1.36 CIs can be used to test hypotheses (does the CI include 0?) CIs also provide a direct estimate of the population means
t-tests
t-tests are used to test the difference between two
means
One-sample t-test: Compare to population mean or
baseline Independent samples t-test: Compare means for two independent groups Related samples or paired t-test: Compare means for two related groups (matched subjects or repeated measures)
Error variance is used to estimate how much the
means of the conditions would differ due to error variance alone. The observed difference between the means is compared with this estimate.
Hypotheses
Directional hypothesis states which of the two
means are expected to differ but does not specify which will be larger
Use a two-tailed test