Block 2
Block 2
Parametric and
Non-Parametric
Statistics
BLOCK 2
PARAMETRIC STATISTICS
43
Inferential
Statistics: An
Introduction
44
Test of
UNIT 3 TEST OF SIGNIFICANCE OF Significance of
Difference
DIFFERENCE BETWEEN TWO Between Two
MEANS* Means
Structure
3.0 Objectives
3.1 Introduction
3.2 Need and Importance of the Significance of the Difference between
Means
3.3 Fundamental Concepts in Determining the Significance of the
Difference between Means
3.3.1 Null Hypothesis
3.3.2 Standard Error
3.3.3 Degrees of Freedom (df)
3.3.4 Level of Significance
3.3.5 Two Tailed and One Tailed Tests of Significance
3.3.6 Errors in Testing
3.4 Methods to Test the Significance of Difference between the Means of
Two Independent Groups t-test
3.4.1 Testing Significance of Difference between Uncorrelated or Independent Means
3.5 Significance of the Difference Between two Correlated Means
3.5.1 The Single Group Method
3.5.2 Difference Method
3.5.3 The Method of Equivalent Groups
3.5.4 Matching by Pairs
3.5.5 Groups Matched for Mean and Standard Deviation
3.6 Let Us Sum Up
3.7 References
3.8 Key Words
3.9 Answers to Check Your Progress
3.10 Unit End Questions
3.0 OBJECTIVES
After reading this unit, you will be able to:
discuss the need and importance of the significance of the difference
between means;
*
Dr. Vivek Belhekar, Faculty, Department of Psychology, Bombay University, Mumbai
45
Parametric
Statistics
explain what is null hypothesis, standard error, degrees of freedom, level
of significance, two tailed and one tailed test of significance, type I error
and type II error; and
calculate the significance of difference between mean ( t-test) when
groups are independent, when there are correlated groups, groups
matched by pair and groups matched by mean and standarddeviation.
3.1 INTRODUCTION
In psychology some times we are interested in research questions like Do the
AIDS patients who are given the drug AZT show higher T-4 blood cell
counts than patients who are not given that drug? Is the error rate of typist the
same when work is done in a noisy environment as in a quiet one? Whether
lecture method of teaching is more effective than discussion method?
Consider the question whether lecture method of teaching is more effective
than discussion method. For this investigator divides the class in to two
groups. One group is taught by lecture method and other by discussion
method. After a few months researchers administer an achievement test for
both the groups and find out the mean achievement scores of the two groups
say, M1 and M2. The difference between these two mean is then calculated.
Now the questions is whether the difference is a valid difference or it is
because of sampling fluctuation or error of sampling. Whether this difference
is significant or not significant. Whether on the basis of this difference, could
we conclude that one method of teaching is more effective than the other
method.
These questions can be answered by the statistical measures which we are
going to discuss in this unit. To test the significance of difference between
mean we can use either the t-test or z test. When the sample size is large, we
employ z test and when sample is small, then we use the t test. In this unit we
are concerned with t test. We will get acquainted with the various concepts
related, to computation and description of t test.
46
Let us take the first question on linguistic ability of boys and girls. First we Test of
Significance of
randomly select a large sample of boys and girls (large sample means the Difference
group comprises of 30 or more than 30 persons.). Then we administer a Between Two
Means
battery of verbal test to measure the linguistic ability of the two groups and
compute the mean scores on linguistic ability test of the two groups. Let us
say the obtained mean scores for boys and girls are M1and M2 respectively.
Now we try to find the difference between the two means. If we get a large
difference (M1 – M2) in favour of the girls then we can confidently say that
girls of 10 years of age are significantly more able linguistically than 10
years old boys. On the contrary if we find small difference between two
means then we would conclude that ten years old girls and boys do not differ
in linguistic ability.
The standard error of the mean can be calculated by the following formula:
SEm or σm = σ/√N
Where
σ= The standard deviation of the sample mean
48 N= The number of cases in the sample.
If the standard error of measurement is large it shows considerable sampling Test of
error. Significance of
Difference
Between Two
3.3.3 Degrees of Freedom (df) Means
Degrees of freedom varies with the nature of the population and the
restriction imposed. For example in the case of value calculated between
means of two independent variables, where we need to compute deviation
from two means, the number of restrictions imposed goes up to two
consequently df is (N- 1+N-2).
In social sciences .05 and .01 level of significance are most often used. When
we decide to use .05 or 5% level of significance for rejecting a null
hypothesis it means that the chances are 95 out of 100 that is not true and
only 5 chances out of 100 the difference is a true difference.
In certain types of data, the researcher may prefer to make it more exact and
use .01 or 1% level of significance. If hypothesis is rejected at this level it
shows the chances are 99 out of 100 that the hypothesis is not true and only 1
chance out of 100 it is true. The level of significance which the researcher
will accept should be set by researcher before collecting the data.
A type I error, also known as alpha error, is committed when null hypothesis
(Ho) is rejected when in fact it is true. A type II error, also known as beta
error, is committed when null hypothesis is retained and in fact it is not true.
For example suppose that the difference between two population means (μ1-
μ2) is actually zero, and if our test of significance when applied to the sample
mean shows that the difference in population mean is significant we make a
type I error. On the other hand if there is true difference between two
population mean, and our test of significance based on the sample mean
shows that the difference in population mean is “ not significant‟ we commit
a type II error.
1) Given below are some statements. Indicate whether the statement is true
or false:
Calculations
M = ΣX/ N = 190/ 6 = 31.66
X M = ΣY/ N = 240/ 10 = 24
Y
SD= √ [( ∑x²+∑y²)/(N1-1)+(N2-1)]
SD = √ [( 110+352)/(6-1)+(10-1)]
= √ 462/14 = √ 33 = 5.74
SE = SD [√( (N1+N2)/(N1N2)]
d
Referring to the critical value in table for t test (that are given in the appendix
of a textbook on statistics) for df= 14, we get 2.58 at the .05 and 2.58 at
the.01 level, since our t is less than 2.14, therefore we will say that the mean
difference between boys and girls is non significant.
Mean SD N df
Boys 40.39 8.69 31 30
Girls 35.81 8.33 42 41
52
Is the mean difference in favour of boys and significant at the .05 level. Test of
Significance of
First we will compute the pooled standard deviation by the following formula Difference
Between Two
SD = √ [(SD1)2× (N1-1) + (SD2)2× (N2-1)]/(N1-1)+(N2-1) Means
SE = SD [√(N1+N2)/(N1xN2)]
d
Referring to the critical value in table for t test (that are given in the appendix
of a textbook on statistics) for df= 71, we find t entries of 2.00 at the 0.05
level and 2.65 at the 0.01 level. The obtained t of 2.26 is significant at .05
level but not at the 0.01 level. We may say boys academic achievement is
better in comparison to that girls.
Some time we have single group and we administer the same test twice. For
example if we are intending to find out the effect of training on the students‟
educational achievement, first we take the measures of participants’
educational achievement before training , then we introduce the training
programme, and again we take the measure of educational achievement. In
this we have single group and administer educational achievement test twice.
Such type of design is known as single group design. In order to get the
significance of the difference between the means obtained in the before
training and after training we use the following method.
Referring to the critical value in table for t test (that are given in the appendix
of a textbook on statistics) for df= 99, and find that the value at 0.05 level is
1.98 and at 0.01 level is 2.63. Our obtained value is 2.35 therefore this value
is significant at 0.05 level and not on 0.01 level. Here we can say that class
made substantial improvement in mathematical achievement in six months.
Note: The sum of scores of final condition is more than sum of initial
condition therefore we subtract scores of initial condition from scores of final
condition (Final condition –Initial condition) and add the score to find ΣD.
Mean = (∑D)/N
= 96/12=8
2
SDD= √[(∑x )/N-1]
=√[354/11]
=5.67
SEMD= √
= 5.67 √
= 1.64
t= MD-0/SEMD
=8-0/1.64
=4.88
df=12-1
= 11 55
Parametric
Statistics
Referring to the critical value in table for t test (that are given in the appendix
of a textbook on statistics) for df = 11, we find t values of 2.20 and 3.11 at
the 0.05 and at the 0.01 levels. Our t of 4.88 is far above the 0.01 level. We
can conclude that participants attitude changed significantly from initial to
final condition.
Here,
σM1= Standard error of mean 1
σM2= Standard error of mean 2
Example: There are two groups X and Y of Children .72 in each group are
paired child to child for age and intelligence. Both groups were given group
intelligence scale and scores were obtained. After three weeks experimental
group participants were praised for their performance and urged to try better.
The control groups did not get the incentive. Group intelligence scale again
was administered on the groups. The data obtained were as follows. Did the
praise affect the performance of the group or is there a significant difference
between the two groups.
56
Table 3.5: Results of Experimental and Control Groups. Test of
Significance of
Difference
Experimental Control group Between Two
Means
group
No. of children in each group 72 72
Mean score of final test 88.63 83.24
SD of final test 24.36 21.62
Standard error of the mean of final test 2.89 2.57
t=(M1-M2)-O/SED = 6.11
t= (88.63-83.24)-0/6.11
=0.88
df = 72-1
= 71
Referring to the critical value in table for t test (that are given in the appendix
of a textbook on statistics) for df = 71, the value at 0.05 is 2.00 and at 0.01 is
2.65. The obtained t is 0-88 therefore this value is not significant at 0.05 level
and not at 0.01 level. On the basis of the results is can be said that praise did
not have significant effect in stimulating the performance of children.
Academic Technical
No. of children in each group 58 72
Mean on Intelligence GTest (Y) 102.50 102.80
SD on Intelligence Test Y 33.65 31.62
Mean on Mechanical Aptitude (X) 48.52 53.51
SD on Mechanical Aptitude X 10.60 15.36
r2 .50
= 4.46
t = (M1-M2)-0/ SED
t = (53.61-48.52)-0/4.46
= 0.63
df=(N1-1)+(N2-1)
=(58-1)+(72-1)
=128
Referring to the critical value in table for t test (that are given in the appendix
of a textbook on statistics) for df = 125(which is near to 128), the value are
1.98 at .05 level and 2.62 at 0.01 our obtained value is 0.63, which is not
significant at 0.05 level. We may say that two group do not differ in
mechanical aptitude.
1) Ten persons are tested before and after the experimental procedure, their
scores are given below. Test the hypothesis that there is nochange.
Before After
60 72
52 50
61 71
36 45
45 40
52 62
70 78
58
Test of
51 61 Significance of
Difference
80 94 Between Two
Means
65 73
72 82
3.7 REFERENCES
Kerlinger, Fred N. (1963). Foundation of Behavioural Research, (2nd Indian
reprint) Surjeet Publication, New Delhi.
Garrett, H.E. (1971), Statistics in Psychology & Education, Bombay, Vakils,
Seffer & Simoss Ltd.
Guilford, J.P. (1973), Fundamental Statistics in Psychology & Education,
Newyork, McGraw Hill.
1) Ten persons are tested before and after the experimental procedure, their
scores are given below. Test the hypothesis that there is nochange.
Before After
60 72
52 50
61 71
36 45
45 40
52 62
70 78
51 61
80 94
65 73
72 82
61
Parametric
Statistics UNIT 4 TEST OF SIGNIFICANCE OF
DIFFERENCE BETWEEN MORE
THAN TWO MEANS*
Structure
4.0 Objectives
4.1 Introduction
4.2 Concept of Variance
4.3 Concept of Analysis of Variance (ANOVA)
4.4 Computation of One Way Analysis of Variance (ANOVA)
4.5 Factorial Design
4.6 Computation of Two Way Analysis of Variance (ANOVA)
4.7 Let Us Sum Up
4.8 References
4.9 Key Words
4.10 Answers to Check Your Progress
4.11 Unit End Questions
4.0 OBJECTIVES
After going through this unit, you will be able to:
4.1 INTRODUCTION
A researcher wanted to study the effect of stages of adolescence, early,
middle and late, on the emotional intelligence of adolescents in Chennai. For
the purpose, she used the standardised psychological tests and collected data
from a sample of 600 adolescents in Chennai, 200 each of early adolescents,
middle adolescents and late adolescents. As the data was organised, the
researcher then had to decide about which statistical technique to use for the
purpose of data analysis. Initially the researcher felt that t- test can be used.
But in the present case, there were three groups and as such t test is used
when there are more than two groups. The research then decided to use one
way ANOVA. Though it would be possible to use t test in this situation as
*
Prof. Suhas Shetgovekar, Faculty, Discipline of Psychology, IGNOU, Delhi
62
well, but then the researcher will have to first compare early and middle Test of
Significance of
adolescents, then early and late adolescents and then middle and late Difference
adolescents with regard to emotional intelligence. This will become further Between More
Than Two Means
cumbersome if there are more than three groups, for instance ten groups and
so on. Also, t test will not provide any information about the variance that
may exist from the mean values of the given groups. In such a situation, one
way ANOVA can be conveniently and effectively used.
In the present unit thus, we will mainly focus on the concept of variance and
then we will discuss the computation of one way ANOVA and two way
ANOVA. In this context we need to remember that both these techniques fall
under parametric statistics and thus the assumptions of parametric statistics
need to be met before these techniques are used for calculations.
The term variance was used to describe the square of the standard deviation
by R.A. Fisher in 1913. The concept of variance is of great importance in
advanced work where it is possible to split the total into several parts, each
attributable to one of the factors causing variations in their original series.
Variance is a measure of the dispersion of a set of data points around their
mean value. It is a mathematical expectation of the average squared
deviations. It mainly helps in separating the variability in to factors that are
random and factors that are systematic (Veraraghavan and Shetgovekar,
2016).
The variance (s2) or mean square (MS) is the arithmetic mean of the squared
deviations of individual scores from their means. In other words, it is the
mean of the squared deviation of scores.Variance is expressed as V = SD2.
The variance and the closely related standard deviation are measures that
indicate how the scores are spread out in a distribution. In other words, they
are measures of variability. The variance is computed as the average squared
deviation of each number from its mean.
In this context we also discuss about between group variance and within
group variance. Between groups variance can be explained as variance that
exists between the group means. For example, the variance in group means of
early, middle and late adolescents. Whereas, within group variance can be
explained as variance that exists amongst the members within a certain
group. For example, variance that may exist amongst the early adolescents. 63
Parametric
Statistics
Let us now focus on some of the characteristics of variance:
1) Variance can be termed as a measure of variability and it provides
information about variance in the scores in similar way as other measures
of variability. You may recall that we discussed about variance in the
context of measures of variability in BPCC104.
2) Variance is denoted in terms of an area. In normal probability curve, the
variance is denoted in terms of area either on the right or left side of the
curve. On the other hand, standard deviation is denoted by direction on
the normal probability curve.
3) The value of variance is always positive.
4) The variance will remain the same, even if a certain constant in a data is
subtracted or added.
The variance (s2) or mean square (MS) is the arithmetic mean of the squared
deviations of individual scores from their means. In other words, it is the
mean of the squared deviation of scores.Variance is expressed as V = SD2.
The formula thus is as follows:
Let us us now focus on the steps involved in computation of one way ANOVA.
Let us compute Sum of Squares (SSt) with the help of the formula
67
Parametric
Statistics
SSt =[(ΣX12 + ΣX22 + ΣX32) – C]
= [(72+ 162 + 88) – 282.13]
= 322- 282.13
= 39.87
Thus, SStis obtained as 39.87.
Step 4: The between groups Sum of Squares ( SSb) is to be computed
The between groups Sum of Squares ( SSb) is computed by squaring the total
of each group divided by respective number of cases, N1, N2, N3 and
subtracting Correction sum (C) from the obtained value.
SSb = (X12 /N1) + (X22 /N2)+( (X32 /N3) - C
= 26 x 26 +38 x 38 + 28 x 28 - 282.13
10 10 10
= (67.6 + 14.44 + 78.4) – 282.13
=290.4 - 282.13
= 8.27
Thus, SSbis obtained as 8.27.
Step 5: Sum of Squares (SSw) is to be computed
Sum of Squares (SSw) is to computed by subtracting the Between sum of
Square from the Total Sum of Square.
SSw = SSt - SSb
= 39.87 – 8.27
= 31.6
Thus, SSw is obtained as 31.6.
Step 6: The degree of freedom (df) are worked out.
For total Sum of Squares N- 1= 30-1= 29
For between group Sum of Squares K- 1 = 3-1 = 2
For within group Sum of Squares N- K = 29- 2 = 27
Step 7: Computation of F ratio
Table 4.3: Summary of ANOVA
68
Thus, F = MSS between groups / MSS within groups Test of
Significance of
F = 4.14/ 1.17 Difference
Between More
F= 3.54 Than Two Means
The F ratio is obtained for the degree of freedom (df) = 2,27
Step 8: Interpretation
Now we have computed the F ratio, but then the value as such is
meaningless, until and unless it is interpreted. For the purpose of
interpretation, you need to refer to certain tables. These tables are given in the
appendix of any Statistics books. The title of the table could be “F ratio for
0.01 and 0.05 levels of significance” (approximately) and may vary to some
extent in different books. In such a table, the degree of freedom for greater
mean square is given as headings for the column (on top) and degree of
freedom for smaller mean square is given as headings for the rows (left hand
side).
Based on the df smaller mean square and df for greater mean square, you can
identify the critical value given in the table for 0.01 and 0.05 levels of
significance.
In the case of our example, the df for smaller mean square as 2 and df for
greater mean square as 27, the critical value is 19.50 at 0.05 level of
significance and 99.50 at 0.01 level of significance. The F ratio obtained in
the example 3.54 is thus not significant as the value is less than the critical
values at 0.05 and 0.01 levels of significance. Thus, it can be said that there is
no significance difference between low, middle and high SES with regard to
achievement motivation.
…………………………………………………………………………….
…………………………………………………………………………….
…………………………………………………………………………….
The design that is used here can be termed as factorial design. Let us discuss the
same before we go on to computation of two way ANOVA.
69
Parametric
Statistics
Factorial designs are mainly used to study the effectof more than two
independent variables on the dependent variable. The main effect (of each
variable separately) as well as interaction effect (of all the IVs) studied with
the help of this design.
Various types of factorial design are as follows:
2 x 2 factorial design: Here there are two independent variables, each
with two categories or levels. For example, gender (male and females)
and managers (junior and senior).
2 x 3 factorial design:Here there are two independent variables. One
with two categories or levels and the other with three categories or
levels. For example, gender (male and females) and Socio Economic
Status (high, middle and low).
3 x 3 factorial design:Here there are two independent variables. each
with three categories or levels. For example, Socio Economic Status
(high, middle and low) and stages of adolescence (early, middle and
late)
n x k factorial design: Here there are two independent variables with n
and k categories or levels. n and k can take any number.
2 x 2 x 2 factorial design: Here there are three independent variables,
each with two categories or levels. For example, gender (males and
females), managers (junior and senior) and Socio Economic Status (high
and low).
3 x 3 x 3 factorial design: Here there are three independent variables,
each with three categories or levels. For example, Socio Economic Status
(high, middle and low), phases of adolescents (early, middle and late)
and religion (Hindus, Muslims and Christians).
n x k x l factorial design: Here there are three independent variables
with n and k categories or levels. n, k and l can take any number.
Two way ANOVA can help in studying the main effect and the interaction
effect of the independent variables on the dependent variables. Main effect
can be described as “ the mean difference amongst the levels of one factor”
(Mohanty and Misra, 2016, page 572). Factor means independent variable.
Thus, when we display the independent variables, they would be form of
columns and rows (as can be seen in table 4.4). Thus, the mean different
between the rows would denote the main effect of one variable and mean
difference in columns will denote the mean difference of the other variable.
70
Table 4.4: Hypothetical example for two way ANOVA Test of
Significance of
Difference
Stages of adolescence (Independent Variable B) Between More
Than Two Means
Gender Early Middle Late
(Independent adolescence (b1) adolescence adolescence
variable A) (b2) (b3)
Male (a1) Mean for a1 and Mean a1and b2= Meana1and b3= Meana1=
b1= 60 55 70 61.7
Female (a2) Mean a2 and b1= Mean a2 and b3= Mean a2and b3= Meana2=
40 65 80 61.7
Mean b1= 50 Mean b2= 60 Meanb3= 75 61.7
As can be seen in table 4.4, the two independent variables are Stages of
adolescence and gender. And in the cells we have mentioned means that
would be based on the dependent variables. The mean difference between
a1and a2 will constitute the main effect for the independent variable A and the
mean difference between b1 and b2 and b3 will constitute the main effect for
independent variable A. Besides main effect, we also need to discuss about
the interaction effect. This is the effect that interaction between variable A
and B has on the dependent variable.
Two way ANOVA can also be effectively displayed in form of line graph.
The interaction effect in this regard can be of three types, that are discussed
as follows:
High SES
Low SES
Males Females
Fig 4.1: Parallel
71
Parametric
Statistics • Additive: Here there is some interaction between the independent
variables with regard their effect on dependent variables. In this, as you
can see in figure 4.2, the lines are not parallel but as such, there is no
complete interaction.
High SES
Low SES
Males Females
High SES
Low SES
Males Females
72
Note: The example discussed here is hypothetical and may not hold true in Test of
Significance of
reality, but has been used in the context for clarity of the concept. Difference
Between More
Check Your Progress IV Than Two Means
Here it is not possible for the researcher to carry out a one way ANOVA as there
are two independent variables that are categorical and the researcher may also be
interested in finding out the interaction between the two independent variables
with regard to their effect on the dependent variables. Thus in this case, the two
way ANOVA can be used.
As you can see, there are a total of six groups in the above example, which are as
follows:
- early adolescence who are males
- middle adolescence who are males
- late adolescence who are males
- early adolescence who are females
- middle adolescence who are females
- late adolescence who are females
In table 4.5, we have calculated n, ΣX, mean and ΣX2 for both rows and
columns.
Let us now try to understand the computation of ANOVA stepwise
Step 1: Compute the correction term (C) with the help of the formula
C = (ΣX)2/ N
In this formula N = n1 + n2 + n3 =10+ 10+ 10= 30.
Thus, N = 30
C = (ΣX)2/ N
74
= (92)2/ 30 Test of
Significance of
Difference
= 8464/ 30 Between More
Than Two Means
= 282.13
Thus, C is obtained as 282.13.
Step 2: Compute total Sum of Squares (SSt)
Let us compute total Sum of Squares (SSt) with the help of the formula
SSt =ΣΣX2 – C
= 322-282.13
= 39.7
Thus, SSt is obtained as 39.7
Step 3: Between group Sum of Squares ( SSbetween) is to be computed.
Between group Sum of Squares (SSbetween) is to be computedwith the help of
the following formula:
SSbetween = Σ (ΣX)2/n - C
= (ΣX1)2+(ΣX2)2+ (ΣX3)2 +(ΣX4)2+(ΣX5)2+(ΣX6)2- C
n1 n2 n3 n4 n5 n6
= (14)2+(21)2+ (16)2+(12)2 +(17)2 +(12)2- 282.13
5 5 5 5 5 5
= 196 + 441+ 256+ 144+ 289+ 144 - 282.13
5 5 5 55 5
= 1470/5 - 282.13
= 294 - 282.13
= 11.87
Thus, SSbetweenis obtained as 11.87
Step 4:Within group Sum of Squares ( SSw) for columns is to be
computed.
Within group Sum of Squares ( SSw) for columns is to be computed with the
help of the following formula:
SSw=SSt- SSb
= 39.7- 12.87
= 27.43
Thus, SSwis obtained as 27.43
Step 5: ‘A’ Sum of Squares (SSa) is to be computed.
This is mainly the main effect of variable A or it can also be mentioned as the
effect of the row. 75
Parametric
Statistics
„A‟ Sum of Squares (SSa) is to computed
SSa = (ΣXa1)2 +(ΣXa2)2 - C
n1 n2
SSa = (51) +(41)2 - 282.13
2
15 15
= 2601 + 1681/15- 282.13
= 4282/15- 282.13
285.47- 282.13
= 3.34
Thus, SSa is obtained as 3.34.
Step 6: ‘B’ Sum of Squares (SSb) is to be computed.
This is mainly the main effect of variable B or it can also be mentioned as the
effect of the column.
„B‟ Sum of Squares (SSb) is to computed
SSb =
Xb1 2 + Xb2 2 + Xb3 2 - C
n1 n2 n3
SSb =
26 2 + 38 2 + 28 2 - 282.13
10 10 10
= 67.6 + 144.4 +78.4 - 282.13
= 290.4 - 282.13
= 8.27
Thus, SSb is obtained as 8.27.
Step 7: AB interactionSum of Squares (SSab) is to be computed.
(SSab) = SSbetween - SSa- SSb
= 11.87 - 3.34- 8.27
= 11.87 - 11.61
= 0.26
Thus, SSab is obtained as 0.26
Step 8: Degree of freedom
In our example, variable A, that is, gender has two levels (represented by
number of rows) and variable B, that is, stages of adolescence has three
levels (represented by number of columns). Thus, r (rows) = 2 and c
(columns) = 3. Further, there are six conditions (k) and the number of
observations in each is 5 (n). The degree of free would be as follows:
df for SSt = dft = N-1 = kn- 1= 30-1= 29
76
df for SSbetween = dfbetween = k-1= 6-1= 5 Test of
Significance of
Difference
df for SSw = dfw = N-k= 30-6= 24 Between More
Than Two Means
The df for SSbetween can be in three parts as given below:
df for SSa = dfa = r-1 = 2-1= 1
df for SSb = dfb = c-1= 3-1= 2
df for SSab = dfab= (r-1) (c-1)= (2-1)(3-1)= 1x2= 2
Step 9: Variance estimate or mean squares (MS) is to be computed.
Variance estimate or mean squares is to be computed with the help of
following formula:
MS = SS/df
We need to compute MS for the following:
MS for variable A
MSa = SSa/dfa
= 3.34/1
= 3.34
MS for variable B
MSb = SSb/dfb
= 8.27/ 2
= 4.14
MS for AB interaction
MSab = SSab/dfab
= 0.26/ 2
= 0.13
MS for within groups
MSw = SSw/dfw
= 27.43/ 24
= 1.14
Step 10: Computation of F ratios
F ratios are computes as follows for variable A, variable B and AB:
Fa= MSa/MSw
= 3.34/ 1.14
=2.93
Fb= MSb/MSw
4.14/ 1.14
77
Parametric
Statistics
=3.63
Fab= MSab/MSw
0.13/ 1.14
=0.11
The summary of two way ANOVA is given in table 4.6
=3.34 =2.93
8.2667/ 2 =3.63
=4.14
=0.13
Consulting such a table the table value for degree of freedom (df) 1, 24 , that
is in the context of variable A, is 4.26 at 0.05 level of significance and 7.82 at
0.01 level of significance. The F ratio for variable A is obtained as 2.93
78 which is less than the table value and thus the F ratio for variable A is not
significant and null hypothesis can be accepted and it can be said that no Test of
Significance of
gender difference exists with regard to perceived parental behaviour. Difference
Between More
With regard to variable B, the F ratio is obtained as 3.63. The df here is 2, 24 Than Two Means
and the table value is 3.40 for 0.05 level of significance and 5.61 for 0.01
level of significance. The F ratio obtained is less than the table value at 0.01
level, but more than the table value at 0.05 level. Thus, it can be said that F
ratio is significant at 0.05 level of significance and it can be said that
significant difference exists in perceived parental behaviour with regard to
stages of adolescence .
4. 7 LET US SUM UP
To sum up, in the present unit we mainly discussed about the concept of
variance. The variance (s2) or mean square (MS) is the arithmetic mean of the
squared deviations of individual scores from their means. In other words, it is
the mean of the squared deviation of scores. Variance is expressed as V =
SD2. The characteristics of variance were also discussed. Further, in the unit
we explained the concept of ANOVA. In 1923, Ronald. A. Fisher reported
about ANOVA, though the name F test was given to it by George W. Snedecor
in honour of Fisher. The assumptions of ANOVA were also described. The unit
then focused on the computation of one way ANOVA with the help of steps and
example. The unit later discussed about factorial designs. Factorial designs are
mainly used to study the effect of more than two independent variable(s) on the
dependent variable. The main effect (of each variable separately) as well as
interaction effect (of all the IVs) studied with the help of this design. Various
types of factorial design were also discussed. Lastly, the unit described the
computation of two way ANOVA with the help of steps and example.
79
Parametric
Statistics 4.8 REFERENCES
King, Bruce. M; Minium, Edward. W. (2008). Statistical Reasoning in the
Behavioural Sciences. Delhi: John Wiley and Sons, Ltd.
Mangal, S. K. (2002). Statistics in Psychology and Education. new Delhi: Phi
Learning Private Limited.
Minium, E. W., King, B. M., & Bear, G. (2001). Statistical Reasoning in
Psychology and Education. Singapore: John-Wiley.
Mohanty, B and Misra, S. (2016). Statistics for Behavioural and Social
Sciences. Delhi: Sage.
Veeraraghavan, V and Shetgovekar, S. (2016). Textbook of Parametric and
Non-parametric Statistics. Delhi: Sage.
80