Chapter 10 The t Test for Two Independent Samples
Transcription
Chapter 10 The t Test for Two Independent Samples
Statistics for the Behavioral Sciences (5 th ed.) Gravetter & Wallnau Chapter 10 The t Test for Two Independent Samples University of Guelph Psychology 3320 — Dr. K. Hennig Winter 2003 Term 1 Once or Twice? Within Subjects Design • • • • • If the subjects are used more than once or matched, this design is called a Within Subjects Design or a Repeated Measures Design. Advantages of Repeated Measures Designs: They take fewer participants. They typically have more statistical power (like a matched t-test). Disadvantages of Repeated Measures Designs: You have to worry about practice effects and carryover effects. We will return to this. 2 1 þ þ þ 3 Between vs. within designs • Between-subjects – comparison of separate groups (men cf. women; ethnicity1 cf. ethnicity2; married cf. unmarried); or matched groups • between-subjects/repeated measures – two data sets from the same sample (patients before therapy cf. after; drug use Grade 9 cf. grade 11; time1 cf. time2) 4 2 Figure 10-2 (p. 310) Do the achievement scores for children taught by method A differ from the scores for children taught by method B? In statistical terms, are the two population means the same or different? Because neither of the two population means is known, it will be necessary to take two samples, one from each population. The first sample will provide information about the mean for the first population, and the second sample will provide information about the second population. 5 Table 10-1 (p. 316) Three major differences: mean difference of the two samples, different ns and thus pooled variance (S p2) 6 3 Logic and procedure: t statistic for independent-measure design • H0: µ1 - µ 2 = 0 • H1: µ1 - µ 2 <> 0 • overall t formula: sample mean − population mean M − µ t= = estimated standard error sM • the independent-measures t uses the difference sample mean difference − population mean difference estimated standard error (M 1 − M 2 ) − ( µ1 − µ2 ) = s( M 1 − M 2 ) 7 t= Reminder • The means of two groups are compared relative to a sample distribution of: – means (“is M 1 different from the population µ?”) SS s2 = – sample variances (proved mean all s2 = σ2) n −1 – mean differences (“is the difference between two sample means, ? M 1-M 2, different?”) 6 2 4 8 Sample n (M = ?) . . Sample 1 (M= ?) 1 2 3 4 5 8 4 9 t-statistic (contd.) • (Review) standard error: measures how accurately the sample statistic represents the population parameter – single sample = the amount of error expected for a sample mean – independent measures formula = amount of error expected when you use the mean difference (M 1-M 2) • Recall: s2 sM = n 2 s( M1 − M 2 ) 2 s s = 1 + 2 n1 n2 • now we want to know the total error, but… 10 5 t-statistic (contd.) • But, what if the size of the two samples is not the same (n1<>n2)? • Recall: SS SS 2 s = n −1 = df • we now have two SS and df values, thus: sp = 2 sp 2 SS1 + SS2 or df1 + df2 ( n1 − 1) s12 + ( n2 − 1) s2 2 = n1 + n2 − 2 11 • Example 10.1 (n = 10 for both groups): A list of 40 pairs of nouns (e.g., dog/bicycle, grass/door) are given two (independent) groups for 5 min.: – group 1 is asked to memorize – group 2 uses mental imagery to aid memorization (e.g., dog riding a bicycle) = a Tx effect group 2 group 1 12 6 Computational procedure t-test for means of two independent samples • Step 1. State hypotheses and select alpha H0: µ1 - µ 2 = 0 H1: µ1 - µ 2 <> 0 α =.05 • Step 2. Determine df df = df1 + df2 =(n1 - 1) + (n2-1) = 9+9 = 18 • Step 3. Obtain data & calculate t 13 (contd.) • Step 3a. For the two samples, A and B, of sizes n1 and n2 respectively, calculate (∑ X 1 ) 2 2 2 ∑ X 1 ∑ X 1 and SS1 = ∑ X 1 − n1 • Step 3b. Estimate the variance of the source population sp = 2 SS1 + SS2 ( n1 − 1) + (n2 − 1) • Step 3c. Estimate the sd of the sampling distribution of sample-mean differences 2 2 s( M1 − M 2 ) = • Step 3d. Calculate t as t= sp n1 + sp n2 ( M 1 − M 2 ) − ( µ1 − µ2 ) s( M 1− M2 ) 14 7 Example (contd.) • Step 3a. • Step 3b. sp = 2 SS1 + SS2 200 + 160 = = 20 ( n1 − 1) + (n2 − 1) 9 +9 • Step 3c. s( M1 − M 2 ) = • Step 3d. t = sp 2 n1 + sp 2 n2 = 20 20 + = 4 =2 10 10 ( M 1 − M 2 ) − ( µ1 − µ 2 ) (25 − 19) − 0 6 = = = 3.00 s( M 1−M 2 ) 2 2 • Step 4. Make a decision. Table lookup df = 18 15 Table look up for df = 18 Obtained value of t is three times greater than would be expected by chance (the standard error): t = 3.00 The Tx moved the mean from M = 22 to M = 25 16 8 Figure 10-4 (p. 318) The t distribution with df = 18. The critical region for α = .05 is shown. Two things are important: significance and effect size 17 Calculating effect size Two methods (Cohen’s d or r2) • Cohen’s d = mean difference/sd = M1 − M 2 s p2 = 25 − 19 6 = = 1.34 4.47 20 • the distance between the two means is slightly more than 1 sd, thus d should be slightly larger than 1.00 • Can also calculate variance accounted for by Tx r2 = t2 32 9 = = = 0. 333 t 2 + df 3 2 + 18 27 18 9 Calculating r2 directly: Tx contribution to variability =Tx variability/ total variability =180/540 = .333 = 33.3% 19 Final comparison 20 10