Computing a one- way ANOVA Rick Balkin, Ph.D., LPC, NCC Department - - PowerPoint PPT Presentation

computing a one way anova
SMART_READER_LITE
LIVE PREVIEW

Computing a one- way ANOVA Rick Balkin, Ph.D., LPC, NCC Department - - PowerPoint PPT Presentation

Computing a one- way ANOVA Rick Balkin, Ph.D., LPC, NCC Department of Counseling Texas A&M University-Commerce Rick_balkin@tamu-commerce.edu Balkin, R. S. (2008). Why do an ANOVA? An Analysis of Variance (ANOVA), also known as an


slide-1
SLIDE 1

Balkin, R. S. (2008).

  • Computing a one-

way ANOVA

Rick Balkin, Ph.D., LPC, NCC Department of Counseling Texas A&M University-Commerce Rick_balkin@tamu-commerce.edu

slide-2
SLIDE 2

Balkin, R. S. (2008).

  • Why do an ANOVA?

An Analysis of Variance (ANOVA), also

known as an F-test, is conducted when two or more group means (J) are being compared.

When J = 2, either a t-test or F-test may

be computed.

The relationship between a t-test and F-

test when J = 2 is F = t2.

slide-3
SLIDE 3

Balkin, R. S. (2008).

  • Why do an ANOVA?

However, when J > 2 (i.e. there are 3 or more

means being compared), statistical significance can be ascertained by conducting one statistical test, ANOVA, or by repeated t-tests.

Why not conduct repeated t-tests? Each

statistical test is conducted with a specified chance of making a type I–error—the alpha level.

slide-4
SLIDE 4

Balkin, R. S. (2008).

  • Why do an ANOVA?

For example, a researcher wishes to use a self-efficacy

screening measure on students in a math class. Research has shown that higher levels of self-efficacy are related to better performance. Students are randomly placed in four groups. Group 1 receives a lecture on enhancing self-efficacy. Group two receives a lecture and experiential exercise on self-efficacy. Group 3 receives an experiential exercise only. Group 4 is a control group. So a self-efficacy measure is administered to four different math classes.

slide-5
SLIDE 5

Balkin, R. S. (2008).

  • Why do an ANOVA?

If t-tests were used to conduct the analysis and

a level of significance was set at .05, then six separate t-tests would need to be conducted: (a) Group 1 to Group 2, (b) Group 1 to Group 3, (c) Group 1 to Group 4, (d) Group 2 to Group 3, (e) Group 2 to Group 4, and (f) Group 3 to Group 4.

  • If J =4, then the number of comparisons is

always

. 6 2 ) 3 ( 4 2 ) 1 ( = = − J J

slide-6
SLIDE 6

Balkin, R. S. (2008).

  • Why do an ANOVA?

The problem with conducting multiple t-tests is that type I error is

multiplied by the number of tests being conducted.

Hence, if a researcher conducts six t-tests with an = .05 on the

same data, the chance of having a type I error among the six tests is 30%, a rather large likelihood of identifying statistically significant differences when none actually exist.

When considering that research can help in identifying what models

may be helpful to clients, one would need to be extremely cautious in implementing best practices when there is a 30% chance of being wrong.

Rather than conducting several t-tests, an ANOVA could be

conducted to identify statistically significant differences among all the groups at the .05 level of significance—only a 5% chance of type I error

slide-7
SLIDE 7

Balkin, R. S. (2008).

  • Computing a one-way ANOVA

To compute the F-test, the example on p.

7 in your notepack will be utilized.

slide-8
SLIDE 8

Balkin, R. S. (2008).

  • Computing a one-way ANOVA

Group1 Group 2 Group 3 Group 4 4 9 8 1 6 11 6 2 8 8 9 3 3 9 5 5 9 8 7 1 6 9 7 2.4 6.1 6.5 1.5 2.5 2.8 =

1

X =

2

X =

3

X =

4

X = . X

=

2 1

s =

2 2

s

=

2 3

s

=

2 4

s

slide-9
SLIDE 9

Balkin, R. S. (2008).

  • State the null hypothesis and alpha

level

Unlike the t-test, the F-test is calculated

using squared deviations.

Remember in the t-test, the numerator

was calculated by subtracting one group mean from another; hence, an observed value for a t-test could be positive or negative and fall on either side of the normal curve.

slide-10
SLIDE 10

Balkin, R. S. (2008).

  • State the null hypothesis and alpha

level

The numerator of the F-test uses squared

values, so the observed value in a F-test is always positive.

Therefore, the curve is somewhat different

slide-11
SLIDE 11

Balkin, R. S. (2008).

  • State the null hypothesis and alpha

level

Therefore, the null hypothesis in a F-test is

never directional. Using the above data set, the null hypothesis and alternative hypothesis would be expressed as follows:

4 3 2 1 1 4 3 2 1

: : µ µ µ µ µ µ µ µ ≠ ≠ ≠ = = = H H o

slide-12
SLIDE 12

Balkin, R. S. (2008).

  • State the null hypothesis and alpha

level

So, in this case, if any of the groups are

not statistically significantly different (either higher or low) from one another, then the null hypothesis is accepted. If a statistically significant difference does exist, then the null hypothesis is rejected.

slide-13
SLIDE 13

Balkin, R. S. (2008).

  • Calculating the F-test:

Understanding the properties

The statistical tests discussed in this class share

common properties. Each test statistic is computed from a fraction. The numerator represents a computation for mean differences, such as by comparing two groups and subtracting one mean from another. The denominator is an error term, computed by using taking in to account the standard deviation or variance, and sample size. The numerator is an expression of differences between groups, often referred to as between-group differences, the denominator looks at error, or differences that exist within in each group, often referred to as within-group differences.

slide-14
SLIDE 14

Balkin, R. S. (2008).

  • The Test Statistic

error s difference mean

error

s

2 1

µ µ −

s difference group within s difference group between − −

  • r
  • r
slide-15
SLIDE 15

Balkin, R. S. (2008).

  • So why does this computation

work?

slide-16
SLIDE 16

Balkin, R. S. (2008).

  • So why does this computation

work?

!"#$%"$$%"&"#

"#

  • $'#
  • (
  • )
slide-17
SLIDE 17

Balkin, R. S. (2008).

  • Understanding the properties

As the computation of the ANOVA is

explained, keep in mind the common properties statistical tests share.

Essentially, ANOVA is calculated by

dividing the mean differences squared by the error variance. The numerator will be mean differences and the denominator will be error variance.

slide-18
SLIDE 18

Balkin, R. S. (2008).

  • Computing the grand mean
  • *

) (

j

X

*

  • .)

(X

*

slide-19
SLIDE 19

Balkin, R. S. (2008).

  • Computing the grand mean
  • !
  • =

j j j

n X n X.

  • +,-.
  • "#/

+

j

X /

+/

slide-20
SLIDE 20

Balkin, R. S. (2008).

  • Computing the grand mean

) ))!

  • 4

3 2 1 4 4 3 3 2 2 1 1

) ( ) ( ) ( ) ( . n n n n X n X n X n X n n X n X

j j j

+ + + + + + = =

  • 1

. 6 20 12 35 45 30 5 5 5 5 ) 4 . 2 ( 5 ) 7 ( 5 ) 9 ( 5 ) 6 ( 5 . = + + + = + + + + + + = X

slide-21
SLIDE 21

Balkin, R. S. (2008).

  • Calculate the sum of squares

between--SSb

As stated before, rather than compare one

group mean to another as was done in a t-test, group means are compared against the grand mean.

The Sum of Squares Between (SSb) is the sum

  • f the differences between each group mean

and the grand mean squared. By comparing each group mean to the mean of the entire sample, the amount of variation between the groups can be assessed.

slide-22
SLIDE 22

Balkin, R. S. (2008).

  • Calculate the sum of squares between--SSb

0!

2

.) ( X X n SS

j j B

− =

  • j

n ,

  • 2

.) ( X X j −

1

.) ( X X j −

slide-23
SLIDE 23

Balkin, R. S. (2008).

  • 122

2 4 4 2 3 3 2 2 2 2 1 1 2

.) ( .) ( .) ( .) ( .) ( X X n X X n X X n X X n X X n SS

j j B

− + − + − + − = − =

0-

2 2 2 2

) 1 . 6 4 . 2 ( 5 ) 1 . 6 7 ( 5 ) 1 . 6 9 ( 5 ) 1 . 6 6 ( 5 − + − + − + −

  • 6

. 114 45 . 68 05 . 4 05 . 42 05 . = + + +

slide-24
SLIDE 24

Balkin, R. S. (2008).

  • 1(*

1(+(/

1 +/

  • 1

3$4

1

3"'

1

slide-25
SLIDE 25

Balkin, R. S. (2008).

  • 1(*
  • 1
slide-26
SLIDE 26

Balkin, R. S. (2008).

  • 1(*

1 5!

) 1 (

2

− Σ =

j j w

n s SS

  • +

2 j

s

/ $+

1 −

j

n

/(!

  • )

1 ( ) 1 ( ) 1 ( ) 1 ( ) 1 (

4 2 4 3 2 3 2 2 2 1 2 1 2

− + − + − + − = − Σ = n s n s n s n s n s SS

j j w

  • 2

. 53 ) 4 ( 8 . 2 ) 4 ( 5 . 2 ) 4 ( 5 . 1 ) 4 ( 5 . 6 = + + + =

w

SS

slide-27
SLIDE 27

Balkin, R. S. (2008).

  • Degrees of Freedom

(

  • 5
slide-28
SLIDE 28

Balkin, R. S. (2008).

  • Degrees of Freedom
  • +/
slide-29
SLIDE 29

Balkin, R. S. (2008).

  • Degrees of Freedom
  • 56785

9

32

132

1+:/ 1+:/

slide-30
SLIDE 30

Balkin, R. S. (2008).

  • 1+:/

:0 $+,2$/!

. 3 1 4 1 = − = − j

  • 2

. 38 3 6 . 114 1 4 6 . 114 1 = = − = − = j SS MS

B B

slide-31
SLIDE 31

Balkin, R. S. (2008).

  • 1+:/

:( +2,/ !;,-"#;<-$4

  • 325

. 3 16 2 . 53 4 20 2 . 53 . = = − = − = j n SS MS

W w

slide-32
SLIDE 32

Balkin, R. S. (2008).

  • Computing the F-ratio

( 32

j n ss j SS MS MS F

w B w B

− − = = . / 1 /

  • 49

. 11 325 . 3 2 . 38 = = F

slide-33
SLIDE 33

Balkin, R. S. (2008).

  • Evaluating the F-test for statistical

significance

3 +&$4/ -$$<'#.

  • 33

50) +,; $/ +; ,/

)

  • $$<'=&"&'
slide-34
SLIDE 34

Balkin, R. S. (2008).

  • Statistically significant results

3.24 11.49