Alemaya Stat
Alemaya Stat
Biometry
FOR
POST GRADUATES
MANDEFRO NIGUSSIE
Ethiopian Institute of Agricultural Research
1
Biometry
PREFACE
Most professionals in a chosen area have been successful due to partly their
understanding of statistics. We can see that most scientists that went up the ladder in
their career understand statistics very well. This does not mean that a brilliant mind in
statistics is a necessary condition for being a brilliant geneticist, breeder, pathologist,
soil scientist, chemist, engineer or any other profession. But it is necessary at least to
design an experiment, present the results of the experiment and understand the real
meaning of any scientific information that backups the results. The fact most
university degrees in natural sciences include a compulsory statistic course is simply
a recognition of this.
Statistics as a science in its own can be very complicated. The statistical methods we
need to use in doing research are only a small and fairly straightforward subset of
this. A general understanding of the basic concepts and principles will be a great
asset in interpreting the experimental results of other people. Understanding this can
give postgraduate students statistical and scientific skills beyond those of the general
public.
There are so many statistical techniques available for different purposes but it is
impossible to master them all in a short period of time. We have therefore selected in
this book of statistical methods that we consider most relevant and useful in
conducting experiments with careful planning, and proper interpretation.
The book is intended for postgraduate students and scientists who have at least a
little background in basic statistics; but seek to learn advanced ideas of statistics and
their application in a variety of practical situations. The core materials of this book are
lecture notes of the course Advanced Statistical Methods (AGR 5201) at Universiti
Putra Malaysia (UPM) and is designed to be covered well within a semester.
We understand that memorizing formulae is of very little practical use and most
computations are carried out by computers these days. However, computers do not
generally tell whether we are carrying out the right computations or not. Here, this
book has a part to play in presenting the necessary procedures needed to do the
analyses either manually or using the software Statistical Analysis System (SAS). We
also presented fundamental principles of experimentation, methods of sampling, data
analysis and proper interpretation of the results that every post graduate student
should possess. In general, the book will provide the important tools needed to be an
effective user of statistics. We hope that it will provide users with most of the
information required for academic as well as for scientific studies in the research
career.
Mandefro Nigussie
2
Biometry
TABLE OF CONTENTS
1. INTRODUCTION
1.1. Importance and Organization
1.2. Basic Statistical Terms and Concepts
2. EXPERIMENTS
2.1. Introduction
2.2. Experimental Designs
2.2.1. Importance of experimental designs
2.2.2. Planning and execution of experiments
2.2.3. Experimental Error
2.2.3.1. Methods of error estimation
2.2.3.2. Methods of controlling error
2.3. Categories of Experiments
2.4. Hypothesis
3
Biometry
4
Biometry
6. FACTORIAL EXPERIMENTS
6.1. Introduction
6.2. Two Factor Experiment
6.2.1. Introduction
6.2.2. Layout
6.2.3. Analysis
6.2.4. Numerical example
6.4. Fractional factorial experiment
8. ANALYSIS OF COVARIANCE
8.1. Introduction
8.2. Covariance Analysis and Blocking
8.3. Numerical example
9. CORRELATION
9.1. Introduction
9.2. Computing r Values
9.3. Correlation and Causation
10. REGRESSION
10.1. Introduction
10.2. Determining a Linear Regression Line with a Single Predictor
10.3. Numerical Example
5
Biometry
1. INTRODUCTION
1.1. Importance and Organization
The core concept underlying all research is its methodology. The research tools are
key elements of the methodology that takes to the ultimate goal of research to
discover something, which was unknown. The physical tools some researchers use
to achieve their goals are distinctly different from those of others, depending on the
discipline. However, the general tools of research that most researchers use to derive
meaningful conclusions are similar. One of the very few important tools of research,
which is common to almost all kinds of research, is statistics.
Statistics play a vital role in virtually all professions. Some familiarity is an essential
component of any higher learning institutions and research organizations. We
understand that all post graduate students and scientific personnel have a good
scientific background in their respective disciplines but their statistical knowledge
ranges from very little to long years of practical experience. Most researchers had
basic statistical courses but they usually face difficulty in applying to their own
practical situations specially in designing sophisticated research work. Hence, this
manual was written to provide researchers with the powerful ideas of modern
statistical methods.
The manual provides basic principles and advanced procedures on the relative
importance of components of variations, the methods to estimate, and means to
control them, in relation to specific experimental designs. The principles and
statistical methodologies illustrated in this manual, provides the type and amount of
data to be collected, the method of organizing data, data analysis and proper
interpretation of the results obtained. It also provides means to draw conclusions,
assess the strength of the conclusions and gauge uncertainties. Examples are drawn
from a wide range of applications to help develop an appreciation of various
statistical methods, their potential uses and their vulnerabilities to misuse.
Statistical concepts have also been explained as an essential component during the
planning stage of an experiment when decisions must be made as to the mode and
extent of the sampling process. This is done because a good design for the process
of data collection permits efficient inferences to be made, with straightforward
analysis.
6
Biometry
What is statistics? A statistical expression that says, ‘we calculate statistics from
statistics by statistics’ is a good example to understand the different uses of statistics.
It is necessary to know the context within which these three uses of the term are
operating. ‘We calculate statistics…’ what it means here is the results of
computations: the means, variances, graphs, tables, correlation coefficients,
regression lines and so on. ‘…from statistics…’ the term statistics here refers to the
raw data to be analyzed. ‘…by statistics.’ The last use of the term is the process or
method that we use to get the desired results and that is the focus of this manual.
The field of statistics has traditionally been divided into two broad categories defined
as: descriptive and inferential. However, these two categories are not mutually
exclusive and that there is a great deal of overlap in what has been labeled as
descriptive may be labeled inferential. Descriptive statistics can be defined as
those methods involving the collection, presentation and characterization of a set of
data in order to properly describe the various features of that set of data (Caulcutt,
1991). The purpose of a descriptive statistic is to tell us something about a particular
group of observations.
Inferential statistics can be defined as those methods that make possible the
estimation of a characteristic of a population or the making of a decision concerning a
population based only on sample results (Walpole et al., 2002). The task of inferential
statistics is to draw inferences or make predictions concerning the population on the
basis of the data from a smaller sample.
The fundamental concepts of statistical inference consist of two major areas known
as parameter estimation and hypothesis testing. Some of the concepts related to
these two areas will be directly used in the development of statistical techniques for
studying experimental designs and the corresponding experimental results. Other
concepts are important prerequisites for a better understanding of the role and scope
of statistical inference in the analysis of experiments.
7
Biometry
completely as possible those critical attributes that distinguish its members from any
and all other entities (Orr, 1995). This step is essential to designate the data or apply
the result. Populations are described by characteristics called parameters and
parameters are fixed values. If only one variable is of interest, the population can be
called as a univariate. For example, the heights of all male students in universities
form a univariate population. Notice that a population is a set of measurements, not
the individuals or objects on which the measurements or counts are made.
The structure of the hypothesis testing will be formulated with the use of the term null
hypothesis. This refers to any hypothesis to be tested and is denoted by H0. The
rejection of H0 leads to the acceptance of an alternate hypothesis, denoted by H1 or
HA. While testing the hypothesis, the critical value is used as a reference point to
differentiate the critical region and the acceptance region. Rejection of the null
hypothesis when it is true is called a Type-I error. Acceptance of the null hypothesis
when it is false is called a Type-II error (Walpole et al., 2002) (See Section 2.4).
Experimental error is the variation between experimental units (plots) treated alike.
There are a number of sources of variations in field experiments, and these sources
contribute to experimental error. These variations could be inherent (natural) or due
to lack of uniformity in the physical conduct of the experiment. One of the goals of
modern experimental designs is to provide a measure of the experimental error and a
means to control the error.
8
Biometry
The experimental error can never be eliminated completely, but can be minimized.
The experimental error can be estimated through replication and randomization and it
can be controlled through blocking, proper plot techniques and application of
appropriate experimental designs. Reduction of the experimental error as much as
possible within the limitations of resources that can be committed to the experiment is
very important for valid comparison of treatment means. The basic idea behind the
reduction of experimental error is that we can detect small differences between
treatment means. Details are given in Section 2.2.3.
9
Biometry
10
Biometry
2. EXPERIMENTS
2.1. Introduction
There are several definitions of the term experiment. Steel and Torrie (1980) defined
experiment as a planned inquiry to obtain new facts or to confirm or deny the results
of the previous experiments. An experiment will always start with an idea, which has
to be formulated into objectives. Objectives are formulated into hypothesis to allow
statistical analysis that will support any conclusion drawn.
Experiment is an important tool for research and should have the following important
characteristics: simplicity, measuring differences at higher precision, absence of
systematic error, wider ranges of conclusion and calculation of degree of uncertainty
(Rees, 1995). The results of experiments are used for decision making like giving
recommendations to use a new product/technology, provided that the new product is
economically more useful than the old version. The new technology (research result)
could vary depending on the field of study. In agriculture, it could be new input
(variety, fertilizer rate/type, pesticide, herbicide, media, feed, vaccination, tree, etc.)
agronomic practices and post harvest management. In education, it could be
teaching method, evaluation technique and so on.
Experimental units and treatments are very general terms. An experimental unit may
be an animal, many animals, a plant, a leaf and so on. Accordingly, a treatment may
be a standard ration, inoculation, and a spraying rate/spraying schedule. In the first
case, the experimenter may compare different rations to identify the best for
recommendation. In the second case, the experimenter may compare different
varieties a crop for disease resistance, and categorize varieties as resistance,
moderately resistance, and susceptible. In the third case the experimenter compares
different rates of herbicide to identify the most effective as well as economical rate
and the right time to apply. In running such experiments, efficient experimental
designs that can provide meaningful answers related to the objectives of the
experiment are needed.
11
Biometry
be obtained from designed experiments that involve statistical procedures for valid
conclusions.
Statement of the problem: In defining the problem, the most important thing is to
describe the scope of the problem, state the major and specific objectives of the
experiment and major assumptions clearly and concisely. It is also necessary to
review what has been done (information or principles established) in the form of past
research achievements, what is desired (goal/purpose of doing the research) and
then analyze the gap (the deference between future desire and past achievements).
Statement of objectives: It takes time and effort to define objectives. Objectives
should be realistic, measurable and achievable. Defining the objectives can be done
in the form of questions to be answered, hypothesis to be tested or effects to be
estimated (Cochran and Cox, 1957; Steel and Torrie, 1980). Some experiments may
have only one objective but others may have a series of specific objectives and in
12
Biometry
such cases they should be written in the order of priority, i.e. the most important one
first, followed by the second most important, and so on.
Choice of experimental site: The choice of experimental site affects the conduct of
an experiment and the interpretation of the results in several ways. It determines the
population to which the results of experiment will apply. Thus the population about
which information is wanted must be defined carefully and the site must be one that
represents the population. For example, if the results are to be used in acid soil
areas, then the experiment should be conducted in a place where soil acidity is a
problem, but not in alkaline soils.
The characteristics of the site also determine the type of experimental design to be
used, the placement of the experiment, the necessity of formation of blocks, and size,
shape and orientation of plots. If the experiment is to be conducted out in the field,
the site could have variation in soil physical and chemical properties, topography, and
so on. The experiment should be planned in such a way that the effects of these
sources of variability are minimized.
Work plan: The easiest and most appropriate form of work plan is the use of a three-
column table that contains type of work in the first column, duration in the second
column and cost attached to each operation in the third column.
13
Biometry
Scope and cost of the experiment: During the planning phase, it is necessary to
consider the scope (local, regional or international) of the experiment, and the cost
attached to it since research costs are usually limited.
When running the experiment, it is vital to monitor the process carefully to ensure that
every thing is being done as planned. Error in experimental procedure at this stage
will usually destroy experimental validity. Up-front planning is crucial to success. It is
easy to underestimate the logistics and planning aspects of running a designed
experiment in a complex and well-organized research system.
Coleman and Montgomery (1993) suggested that prior to conducting the actual
experiment; a few trial runs or pilot runs are often helpful. These runs provide
information about consistency of experimental material, a check on measurement
system, a rough idea of experimental error, and a chance to practice the overall
experimental techniques. This also provides an opportunity to revisit the decision on
the choice of treatments, experimental design and the variable of interest.
Data analysis: Statistical methods should be used to analyze the data so that the
results and conclusions are objective rather than subjective judgmental in nature.
There are many excellent software packages designed to assist in data analysis.
Often we find that simple graphical methods (trend analysis) play an important role in
data analysis and interpretation. Hypothesis testing and confidence interval
estimation procedures are very useful in analyzing data from designed experiments
because many of the questions the researcher wants to answer can be casted into a
hypothesis-testing framework.
All data analysis should be done as planned and the results interpreted in line with
the experimental conditions and the relations of the results to previous findings.
Here, it should be noted that statistics couldn’t improve the variation during the
physical conduct of the experiment. Therefore, consider the consequence of making
incorrect decisions and be reserved to give conclusions against established facts, if
the conclusions appear out of line with previously research results.
The primary advantage of statistical methods is that they add objectivity to the
decision-making process. Statistical techniques, coupled with good scientific out look
(process of knowledge) and common sense will usually lead to a sound conclusion.
Report writing: A complete readable and correct report of the research should be
prepared. Once the data have been analyzed, the researcher must present precise
results and draw practical conclusions about the results and recommend the course
14
Biometry
Steel and Torrie (1980) explained the two sources of variations as inherent variation
and variation due to lack of uniformity in the physical conduct of the experiment.
Inherent variation is due to inconsistency within the population of experimental units.
It cannot be avoided but can be minimized by experimental designs. In general, the
sources of inherent variability in field experiments are plant variability, seasonal
variability and soil variability. Detail examples of the inherent variations include
difference in soil fertility, slope, pH, temperature, sunshine, wind, etc.
It is important that every possible effort be made to reduce the experimental error in
order to improve the power of a test, and to decrease the size of confidence intervals.
This can be accomplished by reducing the two sources of experimental error that is,
reducing the effect of the inherent variability and refining the experimental
techniques. One of the goals of modern experimental design is to provide a measure
of experimental error and to reduce it as much as possible within the limitations of the
resources that can be committed to the experiment.
15
Biometry
In general, a good experimental design minimizes error, so that the treatment effect
can be detected more readily.
I. Replication
To obtain an estimate of experimental error, the scientist must repeat each treatment
two or more times. This repetition is called replication, which is each treatment must
appear more than once in an experiment. The precision of an experiment can always
be increased by having additional replications. Little and Hills (1978) reported that, to
double the precision in an experiment with which two means can be separated, 16
replications are needed as compared to an experiment with four replications.
In general, in field research, three to six replications are enough for reasonable
precision, but it all depends on the resource and degree of precision required.
Cochran and Cox (1957) provided a formula which could conveniently estimate the
number of replications required to detect a specified difference. The formula is
r ≥ 2[(c.v.)2/D2](t1+t2),
s2
SE(m) =
n
2
where s is the sample variance and n is the number of observations (replications)
on which the mean is based. Therefore, as the number of replications (n) is
increased, the standard error of a mean decreases, and as a result, precision
increases.
c. It can broaden the base for making inferences. As replication is increased, a wider
variety of plots can be brought into the experiment. This can lead to a greater
range of conditions over which the results of the experiment can be applied.
16
Biometry
What are the factors affecting the number of replications? There are several
factors that affect the number of replications.
a. Pattern and magnitude of variability in the experimental units: If the experimental
units are more uniform, lower number of replications is required, and vice versa.
b. Size and differences to be detected: In order to visualize minor differences, more
number of replications is required to maximize the degree of freedom for error and
minimize the error mean squares, which in the final analysis, increases the
calculated F-value.
c. Required level of precision: As the number of replications increases, the precision
also increases up to a certain level, any further increase in number of replication
does not add to increase in precision any more.
d. Number of treatments: As the number of treatments is high, the number of
replications can be reduced, in order to have manageable experimental area. This
is true in areas of plant breeding, at the initial stage of progeny generation. Here,
the target is to narrow down the number of genotypes to a manageable size. For
example, we may use two replications to evaluate 200 progenies out in the field.
e. Availability of resources devoted to the experiment: As the number of replication
increases, other resources used in terms of cost, labor and time also increase.
f. Experimental design used: Certain designs require a fixed number of replications.
For example, in a Latin square design with five treatments, it is necessary to use
five replications to analyze the data as a Latin square experiment. The same is
true with a balanced lattice design, which will be explained in more detail in
Section 4.6.2. Other designs like randomized complete blocks can take any
number of replications, but each treatment should appear at least once in each
block. On the other hand, completely randomized designs, any number of
replications can be used and it is not necessary to have all treatments in each
block, i.e. less restricted.
II. Randomization
Randomization is an underlying the use of statistical methods in experimental design
(Snedecor and Cochran, 1980). The major function of randomization is to provide a
valid or unbiased estimate of experimental error and treatment means by ensuring
random assignment of every treatment to any experimental unit. It is one of the few
characteristics of experimental design and involves the use of some chance device
such as use of random number tables. It is also a precaution against disturbances
that may occur while conducting the experiment (like flooding, shade, etc) and it is
done to ensure independence among observations, which is one of the assumptions
for the analysis of variance.
17
Biometry
where adjacent plots tend to be more alike in productivity than plots which are apart.
Such plots are said to give correlated error components or residuals. Hence, the
precision of comparisons among treatments, which are physically close, is greater
than those among treatments, which are some distance apart. Randomization tends
to destroy the correlation among errors and make valid tests of significance (Gomez
and Gomez, 1984).
With proper randomization, any plot is equally likely to be assigned to any treatment.
In some experimental designs, randomization is restricted in certain ways, but in no
design it is completely eliminated. There are two main reasons for randomization: to
eliminate bias and to ensure independence. Randomization assures that no
treatment is favored or handicapped while assigning to the plots.
Faulty techniques may increase the experimental error in two ways (Steel and Torrie,
1980). It may introduce additional fluctuations in a more or less random nature, and
18
Biometry
possibly subject to the law of chance. Such fluctuations reveal themselves in the
estimate of the experimental error, possibly by the coefficient of variation. If
experimenters find that their estimates of experimental error are consistently higher
than those of others in the same field, they should carefully scrutinize their technique
to determine the origin of the errors. The other way in which faulty techniques may
increase experimental error is through nonrandom mistakes. These are not subject to
the law of chance and may not always be detected by observation of individual
measurements. Faulty technique may also result in measurements that are
consistently biased. This does not affect experimental error or differences among
treatment means, but does affect the values of treatment means. Experimental error
cannot detect bias. The experimental error estimates precision or repeatability, not
the accuracy, of measurements.
Cochran and Cox (1957) pointed out that the principal objectives of a good technique
are: to secure uniformity in the application of treatments, to exercise sufficient control
over external influences so that every treatment produces its effects under
comparable and desired conditions, to devise suitable unbiased measures of the
effects of the treatments, and to prevent gross errors, from which no type of
experimentation seems to be entirely free.
d. Concomitant measurements
Whenever inconsistency of the treatment over replications occurs, there is a need to
use concomitant observations. One of the techniques for reducing experimental error
is to remove the variability in the dependent variable (Y) associated with some
independent variable (X). This technique is called covariance. It is the use of the
independent variable as a covariate to adjust the treatment means and the
observation is taken simultaneously with the main observation, but not affected by
the treatment.
A good example in plant breeding may be the evaluation of 10 varieties for maize
grain yield, where the varieties varied in maturity from 70 to 120 days. Here, it is not
faire to compare a variety that can mature within 70 days with another variety having
maturity of 120 days, because maturity has direct relation with yield. The probability
that the late variety would be better in yield than the early one is high. Hence, the
better way of comparison is to consider maturity as covariate and analyze the data to
identify the variety that can give the best grain yield per unit time or resource use.
19
Biometry
In field experiments, long and narrow plots and square blocks provide the greatest
precision (Steel and Torrie, 1980). Little and Hills (1978) also reported that
rectangular plots are most efficient in overcoming soil heterogeneity when their long
axis are in the direction of greatest soil variation.
In general, the design used in conducting the experiment determines the size and
shape of plots and blocks. For some types of experiment, the experimental units are
carefully selected to be as uniform as possible to reduce the experimental error. For
example, in studying the effect of fertilizer, herbicide, fungicide or insecticide, it is
more useful to determine how the experimental units respond to increasing doses of
the treatment than to decide whether or not two succeeding doses are significantly
different.
Faulty techniques could also result from measurements that are consistently biased.
Care and skill in managing measuring devices are important safeguards against such
biases. On top of that, simplified techniques without sacrificing accuracy are a
consideration from time to time.
20
Biometry
across locations and/or years. In this kind of experiment treatment, treatments are
unreplicated. Another example is a uniformity trial planted with the same variety to
understand the direction of soil variability in field experiments.
In a critical experiment, the use of a randomized design is necessary but the choice
of the type of design depends on the objective of the experiment, experimental area
and the number of treatments. In this experiment, treatments are compared and
meaningful differences are detected based on the observations of the responses.
Hence, precision and accuracy are very important elements of critical experiments,
and these are achieved through the use of appropriate experimental designs, which
are illustrated in the sections ahead.
2.4. Hypothesis
Statistical tests start with a statement called a null hypothesis, which is always along
the line ‘there is no difference between treatments’ or ‘there is no relationship
between the measurements’. The structure of the hypothesis testing will be
formulated with the use of the term null hypothesis. The null hypothesis is given a
symbol H0. To test whether the null hypothesis is true or not depends on the
probability level (p value). The p value given by the test tells us the probability of
getting such a result if the null hypothesis is true. A high p value indicates that the
hypothesis could be easily being true and as a result we should not conclude that
there is a difference. A low p value indicates that there is a difference.
The alternative hypothesis is a statement which usually takes the form like ‘there is a
difference between treatments’ and is given the symbol H1 or HA. The two
hypotheses cover all eventualities. A good example is to investigate whether
supplementary irrigation can improve grain yield of maize in a given environment. In
this example, there is a possibility that supplementary irrigation could reduce yield
(like by water logging if there is excess water), otherwise the result of the test will
bias in favor of finding what we hope to find. Therefore, the test should allow for both
possibilities, called two-tailed test (with the two-sided alternative hypothesis). Hence,
the alternative hypothesis is that supplementary irrigation can bring about difference
in grain yield of maize (either increase or decrease). In short, the two hypotheses can
be written as:
21
Biometry
In some cases, the only possibilities are: there is no difference (H0) and one
treatment is greater than the other (HA). Taking the same example, adding
supplementary irrigation in area where the rainfall is not sufficient, always improves
grain yield because adding water where there is water stress does not reduce grain
yield of maize. In such cases we use a one-tailed test (with one-sided alternative
hypothesis).
Hence the alternative hypothesis is that supplementary irrigation increase grain yield
of maize. It is not common to use a one-tailed test in biological science. The most
important thing is that null hypothesis does not change but it is only the alternative
hypothesis that changes into either one- or two-tailed test.
While testing the hypothesis, the critical value is used as a reference point to
differentiate the critical region and the acceptance region. The rejection of H0 leads to
the acceptance of an alternate hypothesis. The error committed in the case of
rejecting the null hypothesis when it is true is called a Type-I error. Error committed
in accepting the null hypothesis when it is false is called a Type-II error (Walpole et
al., 2002).
22
Biometry
A data set may range from a few entries to hundreds of or even thousands of them.
Each entry corresponds to the observation of a specified characteristic of a sampling
unit. For example, if an animal breeder would like to compare different breeds of
dairy cows and record their milk yield per day, then the dairy cows are sampling
units, and the data set would consist of measurements of milk yield per day from the
animals as experimental units. Once the data are collected, a primary step is to
organize the information and extract descriptive summary that highlights its salient
features. The following sub-sections will cover steps to organize and describe a set of
data by means of tables, graphs and calculation of some numerical summary
measures.
23
Biometry
Discrete quantitative data are numerical responses, which arise from a counting
process, while continuous quantitative data are numerical responses, which arise
from a measuring process.
According to Berenson et al. (1988), there are four kinds of scales with different
levels of precision in treating numbers. These are, in increasing order of precision:
nominal, ordinal, interval and ratio scales.
Nominal scale is the simplest and most elementary type of measurement where
numbers are assigned for the sole purpose of differentiating one object from another.
When numbers are used in a nominal scale, it cannot be added them together, or it is
not possible to calculate an average, because the scale does not have the necessary
properties to do so. Examples of nominal scale can be names of highways like High
way 45, Street 54, etc.
Ordinal scale implies the measurement that has the property of order. Here one
object can be differentiated from the other and the direction of the difference can also
be specified. Statements like ‘more than’ or ‘less than’ can be used because the
measuring system has the property of order. The objects or events can be placed on
a continuum in terms of some measurable characteristics. The ranking of objects or
events can be a good example of an ordinal scale. For example, third year students
of Bioindustry: Iden, Nik and Nasir might be ranked as 1, 2 and 3, respectively, based
on their CGPA, but how much difference that exists among them may not be known.
Interval scale is known for its character to have equality of units. There are equal
distances between observation points on the scale. This scale specifies not only the
direction of the difference, as in the ordinal scale, but also indicates the amount of
the difference as well. The measurement that characterizes education and behavioral
science are of the interval type. For example, most test scores, in terms of number of
items corrected, are based on interval measures.
Ratio scale has all the characteristics of interval scale plus an absolute zero. With an
absolute zero point, statements can be made on ratios of two observations, such as
‘twice as long’ or ‘half as fast’. Most physical scales such as time, length and weight
24
Biometry
are ratio scales. For example, if Halima can run 5 kilometers in 30 minutes, but it
takes Siti 60 minutes, then it is known that Halima can run twice as fast.
Sampling is the taking or measuring of more than one observation per experimental
unit. It is not a design but it is an aspect of experimental design. Sampling error
occurs during sample measurements. The sampling error is the variation among
observations taken within experimental units. The main reason for sampling is to
control this variation. It can be done in green houses, under filed conditions, in
laboratories, on life plants like perennial crops, on animals, and so on. Sampling can
be done in one run, at two stages, three stages or so.
The main reason for sampling is to save resources (time, money and efforts). The
second reason for sampling is that, even though part of all the information about the
population is there, the sample data can be useful in drawing conclusions about the
population, with appropriate sampling method and sample size. The third reason for
sampling applies to the special case where the act of measuring the variable
destroys the individual, such as in destructive sampling. Clearly, testing a whole
batch of explosives would be inappropriate, for example.
There are many ways of selecting a sample from a population, but the most important
is random sampling. Indeed, the methods of statistical inference used in this manual
apply only to cases in which the sampling method is random. However, there are
other sampling methods, which are worth mentioning for reasons stated below.
25
Biometry
Where the individuals are objects, which are fixed in a given location, the method of
random sampling can also be used. There are many examples in most fields of study
where such samples are required. One method of obtaining a random sample is to
over lay a plan of the area to be investigated with a rectangular grid, which includes
all points of interest in the area.
The strata may have quite different distribution for the variable of interest. For
example, for the population of heights of maize varieties, it is necessary to separate
early, intermediate and late varieties.
26
Biometry
plants of different strata and this may justify an investigation of possible treatment
stratum interactions.
Let the stems in each selected hills are divided into three (bottom, middle and top)
strata based on the internode lengths on the stems, that is, shortest for the bottom,
intermediate for the top and longest at the middle strata and a random sample of
three internodes will be taken from each stratum, then the total number of sample per
plot will be 3 x 3 x number of hills. In this example stratification was applied on the
secondary unit of the two-stage sampling method, but the stratification can be
applied on the primary unit (stratifying the hills depending on a certain gradient) and
on both primary and secondary units.
27
Biometry
The sequential nature of the sampling process makes it unnecessary to list every
element in the study area. Yet, the random selection at each stage means that every
element has an equal chance of being included in the sample.
The most common question asked by every investigator who whishes to collect and
analyze data is, ‘how much data should be collected?’ Here, it refers to the number of
observations to be included in the sample as the sample size, so the investigator
should be asking, ‘what sample size should be chosen?’ A sample size of 30 sounds
enough, depending on conditions.
28
Biometry
Mead et al. (1993) suggested three possible sample sizes: the first suggestion is to
take 47 units equally from each stratum. The second suggestion is to take 20% of the
units from each stratum (this is called proportional sampling) and the third one is to
take optimal samples from each stratum. The third method would have sample size
proportional to number of samples x standard deviations (nisi). For example, if the
plot to be sampled is stratified into 3 strata, having 200, 400 and 800 number of
samples, then for method-I (equal sampling) 47, 47 and 47 samples will be taken for
each stratum. For method-II (proportional sampling), 40, 80 and 160 will be taken for
each stratum and for method-III (optimal sampling), 40, 160, 80 will be taken for each
stratum. The third method was criticized by Townend (2002) who suggested that the
effect of changing the sample size is to alter the sensitivity or accuracy of statistical
method.
If the cost or effort in sampling is different for the strata then it can be shown that the
sample size should be inversely proportional to the square root of the costs or efforts
involved, as well as to the product of the stratum size and the stratum standard
deviation (Mead et al., 1993).
In designed experiments, simple random sampling is considered and the formula for
the variance to estimate, for example, the total yield of the plot only measures the
sampling precision for the plot. It cannot be used to decide on the number of plants to
be sampled from the plot or it cannot be used to calculate standard errors for
comparing treatment means. These are calculated from analysis of variance of the
estimates of the yield for each plot.
Precision of a sample estimate generally increases with the sample size, the number
of sampling units per plot and the complexity of the sampling design used. But the
increase in any of these three sampling components entails cost that necessitates
the choice of appropriate sampling technique to maintain the balance between
sampling size, size of sampling unit and sampling method to achieve minimum cost.
There is no formula to get the proper combination of these sampling components but
it all depends on the required information, the variability of the character of interest
and the degree of precision to be estimated. In general, data from previous
experiments, ongoing experiments and data from specifically planned sampling
studies can be used as sources of information. In using methods that have been
used in similar way by other workers, a much simpler approach is to follow published
study and draw on their experience. If the intension is to achieve higher precision of a
given results than what has been done, the use of larger sample size is required.
29
Biometry
descriptive statistics have been covered in the basic statistical courses and are also
commonly available in any text.
3.5.1.Frequency distribution
Basically, the frequency distribution is simply a table constructed to show how many
times a given score or group of scores occurred. We can set up a table where the
highest score is at the top and the lowest at the bottom, with all possible scores in
between, and indicate how often each score occurred. The most common form of the
frequency distribution is the grouped frequency distribution (Table 3.1).
Table 3.1. Group frequency distribution of plant height of a sweet corn population
(BC2-10 MSC1)
___________________________________
Class frequency
___________________________________
156-160 10
161-165 80
166-170 235
171-175 370
176-180 220
181-185 80
186-190 5
___________________________________
Source: Mandefro (2004)
Real limits and apparent limits: The apparent limit is like the one written as 156-
160, 161-165, and so on. To preserve the continuity of the measuring system for
calculations to be discussed later on, there should not be a gap between 160 (the top
score in the 156-160 interval) and 156 (the lowest score in the 156-160 interval). For
this reason, it is understood that the real limits of any interval extended from 1/2 unit
below the apparent limit to 1/2 unit above the apparent upper limit. Thus, the real
limits of the 156-160 intervals are 155.5 and 160.5. The real lower limit is designated
L and the real upper limit is U.
Mid point is the exact center of any interval. The mid point of any interval is found by
adding the apparent upper limit to the apparent lower limit and then dividing by 2. For
156-160 interval, the mid point is (156+160)/2 = 158. Interval size is the distance
between the real lower limit and the real upper limit. In other words, it is the distance
between L and U. For the interval 165-160 for example, i is 160.5-155.5 = 5.
The first step in constructing a frequency distribution from any group of data is to
locate the highest and the lowest scores. The distance between the highest and
lowest score is the range. The next step is to determine i, the size of the interval.
Dividing the range by the number of intervals that will be employed does this. Most
commonly i values of 3, 5, 10, 25, 50, and other multiple of 10 are used (Berenson et
al., 1988). The choice of the number of intervals and the size of the interval is quite
30
Biometry
arbitrary. The main point is to have the frequency distribution display as much
information as possible concerning the concentrations and patterns of scores. The
top interval should contain the highest score, and the bottom interval begins with a
multiple of interval size (Caulcutt, 1991).
After determining the size i and the number of intervals to use, then it is a matter of
placing intervals in a column labeled scores at the left side and then begin going
through the data, placing a tally mark by the interval in which each score lies. Each
entry in the frequency column is the addition of these tally marks for each interval.
Many quantitative measurements (height, weight, yield, etc.) are normally distributed
in the population. A normal curve has a bell-shaped symmetrical curve. There will be
times when a graph of the frequency distribution will not have the typical, symmetric
bell-shaped with the majority of scores concentrated at the center of the distribution.
Instead the majority of the scores are clustered at either the high end or the low end
of the distribution. This concentration of scores or measurements at one end or the
other of the distribution is called skewness (Caulcutt, 1991).
If the measurements are concentrated at the upper end of the distribution so that the
tail of the curve skews to the left, we say that the curve is negatively skewed. If the
measurements are clustered at the lower end of the distribution so that the tail of the
curve skews to the right, then the curve is positively skewed. Positive skewness is
characterized by a preponderance of low score, such as plants grown in stressed
environments. Another property of a frequency distribution besides the amount of
symmetry is its kurtosis. Kurtosis is the peakdness or flatness of a frequency
polygon. If the curve has a very sharp peak, it indicates an extreme concentration of
scores about the center, it is said to be leptokurtic. If the curve is quite flat, it would
tell that while there is some degree of concentrations at the center, there are quite a
few measurements that are dispersed away from the middle of the distribution, the
curve is called platykurtic. The curve that represents a medium kurtosis is called a
mesokurtic curve.
3.5.2.Histogram
Histograms are vertical bar charts in which the rectangular bars are constructed at
the boundaries of each class. The steps in graphing frequency distribution are: to
layout an area on a paper having the height of the graph about three-fourths of the
width. The horizontal line, called the x-axis or the abscissa, is drawn long enough to
include all of the scores plus a little unused space at each end. This axis is labeled
and put a number of scores at appropriate intervals. At the left end of the x-axis a
vertical line (the ordinate or x-axis) is drawn. Then the x-axis is divided into units so
that the largest frequency will not reach the top of the graph.
Simply, drawing lines parallel to the x-axis at the height of the frequency for each
interval and connecting the lines to the x-axis by vertical lines to the real limits of the
intervals complete the histogram. The histogram is given a title below the figure and
the title should be a clear statement of what the histogram is representing. When
31
Biometry
comparing two or more sets of data, the various histograms cannot be constructed on
the same graph because superimposing vertical bars of one on another would cause
difficulty in interpretations. For such cases, it is necessary to construct relative
frequency polygons.
3.5.3.Frequency polygon
The polygon is formed by letting the mid point of each class that represent the data in
that class and then connecting the sequence of midpoints. The steps in graphing a
frequency polygon are as follows. First, the area is determined for the graph with the
proper proportions, the x-axis and y-axis are labeled as in histogram. Secondly, a dot
is placed above the mid-point of each interval. After placing the points for all intervals,
the points are connected with straight lines. Thirdly, the curve connecting the points
drops down to the base line at the extreme ends of distribution because it is not
impressive to leave the polygon suspended in midair. The curve touches the base
line at the mid-point of the adjacent interval, whose frequency is zero.
Polygons provides with a useful aid for comparing two or more sets of data. The
properties of central tendency, dispersion and shape can be depicted by comparing
the polygons of particular distributions of data, like normal distribution. Central
tendency and dispersion will be discussed in sections ahead.
The central tendency, dispersion and shape are three major properties that describe
a set of numerical quantitative data. In data analysis and interpretation, a variety of
descriptive measures representing the properties of central tendency, dispersion and
shape may be used to extract and summarize the salient features of the data set
(Rees, 1995).
Mean is the arithmetic average of the values. To calculate the mean, all
measurements are added and then be divided by the number of observations.
32
Biometry
=
∑ xi
n
If for example, yield per plot of five treatments is 150, 200, 250 240 and 260 kilogram
(kg), then the mean yield of the five treatments will be 220 kg, i.e., 150 + 200 + 250 +
240 + 260)/5 = 220 kg. Statistically this can be rewritten as:
x =
∑ xi
n
1100 kg
=
5
= 220 kg, where Σxi is the sum of x-values and n is the No. of values.
In short, a value 220 kg is the best single value that describes yield distribution.
Another useful feature of the mean is that it makes it possible to compare an
individual performance with the rest of the group. A single yield per plot that can be
below or above the mean can be determined. Then, how far below or above the
mean can be seen in terms of deviation or distance from the mean. The deviation can
be obtained by subtracting the mean from each value.
For example, in comparing 10 genotypes, ear height (cm) data were collected from
40 plants for a genotype (10 plants/replication for four replications) and the data were
grouped as in Table 3.2. In the Table 3.2., the first two columns (scores and
frequency) are the usual classes of frequency distribution. The third column, fx, is
simply the result of multiplying each score by its f value. Hence, the mean is equal to
65.425 cm, i.e., the sum of fx divided by the number of observations, n.
33
Biometry
67 3 201
66 6 396
65 5 325
64 3 192
63 6 378
62 5 310
61 1 61
_____________________________________
Total 40 2617
_____________________________________
Source: Mandefro (2004)
Median is the value that exactly separates the upper half of the distribution from the
lower half. It is a measure of central tendency in that the median is the point located
in such a way that 50% of the scores are lower than the median and the other 50%
are greater than the median. It is important to not that the mean is the exact center of
the deviations of the scores from the mean whereas the median is the exact center of
the scores themselves. The calculation of the median requires arrangement of the set
of scores in the order of magnitude, usually the highest score on the top. The median
is then found by counting up from the bottom (n+1)/2 scores. The procedure is
slightly different depending on whether the number of scores is odd or even.
When n is odd, the median is obtained by counting up from the bottom (n+1)/2. For
example, for 11 observations (11+1)/2 = 6 that is the 6th observation is the median. If
n is even, the same formula can be used, suppose there are eight observations
(8+1)/2 is 4.5. This means that the median is half way between the fourth and the fifth
scores from the bottom. This would be the average of the two values (value 4 + value
5)/2. Here, the median is not of the actual values in the distribution but by definition,
one-half of the observations are above the median and half are below. The median is
a valuable measure of central tendency for measurements that are only of the ordinal
type.
In summary, mean is the most stable of the three measures of central tendency. That
is stability in repeated sampling the means of the samples will tend to vary the least
among them. If a constant is added to each observation then mean will be increased
by the amount of the constant. In negative skewness, the mean is the lowest value of
the three measures of central tendency. In positive skewness, the mean is the
highest value of the three measures of central tendency. Therefore, median is the
34
Biometry
Range is one of the simplest and most strait forward measures of variability. It is the
difference between the highest and lowest scores in the distribution. If the highest
weight of sweet corn is 10 kg per plot and the lowest weight is 5 kg per plot, then the
range will be 10 - 5 = 5 kg per plot.
Range has two serious weaknesses. First, one extreme value can greatly alter the
range. These extreme values are called outliers. The outliers could affect the
outcome of our data analysis. The second weakness is that, since the range is based
on only two measures, the highest and the lowest, it tells us nothing about the pattern
of the distribution. It is this last point that is important in attempting to measure
variability in an accurate way.
If variability is defined as the fluctuation of the scores about the central point, a
method that will tell how much those scores are fluctuating about that point, or how
far they are from the mean is required. That will lead to look at the next method that
helps measure the amount of deviation.
Average deviation is a measure of variability that takes into account how far each
score deviate from the mean, the deviation score (x- x ) would give us this
information. If the values of (x- x ) relatively small, it would indicate that there is less
variability and the scores are near the mean. If on the other hand the values of (x- x )
are large then the scores must be scattered farther from the mean. To get a measure
of variability, it would be necessary only to add the (x- x ) values that give an overall
picture of how much variation there exist.
The value of the average deviation would tell that on average each observation
deviates by that value from the mean. Therefore, the average deviation is a sensible,
35
Biometry
Standard deviation is the most widely used measure of variability (Sokal and Rohlf,
1987). This statistic makes use of the deviation of each score from the mean, but the
calculation, instead of taking the absolute value of each deviation, squares each
deviation to obtain values that are all positive in sign. When each of the deviations
are squared, positive numbers are obtained whether the original deviations were
positive or negative because when two numbers of the same sign are multiplied
together their product is always positive. The formula for standard deviation (s) is
s=
∑ (x - x ) 2
n -1
What does s tell? It tells in a relative fashion how many of the scores in a distribution
deviate from the mean. If s is small, there is little variability, and the majority of the
observations are tightly clustered above and below the mean. If s is large, the scores
are more widely scattered about the mean. One of the primary uses of s is to
compare two or more distributions with respect to their variability. Standard deviation
is sensitive to the pattern of score in a distribution. The scores having s value of 4.2
are more tightly clustered than scores having s value of 6.9 for another group.
Note that adding a constant number to each score does not affect s (remains
unchanged) but it alters the mean value by the amount equal to the constant. In
computing s using the computational formula, a negative number may be obtained
under the square root. In such cases, a mistake was made somewhere and hence it
is necessary to recheck the work.
s2 =
∑ ( xi − x)2 .
n −1
The sample mean ( x ) and population mean (µ) have been introduced earlier. Here,
the symbol σ (sigma, lower case Greek letter) and its square (σ2) represent the
population standard deviation and variance, respectively. The formula for σ2 is given
below.
2 ∑ ( x − µ )2
σ = .
N
For population variance, the denominator n-1, which was used for sample variance
cannot be used for population because all the elements of the population are used in
the computation.
Standard
Standard deviation and the normal curve
36
Biometry
In a normal curve, the mean is drawn from the baseline that divides the distribution
into two equal parts. In other words 50% of the scores lie below the mean (to the left)
and 50% above the mean (to the right). Vertical lines can also be drawn from the
base line corresponding to the different σ units so that the area under the curve
between 1σ below mean and 1 σ above the mean is approximately 68% of the total
area. Similarly, about 95% of the distribution lie between -2 and +2 σ units from the
mean, and about 99% of the distribution lie between -3 and +3 σ units from the mean
(Wardlaw, 2000).
The z scores is a way of telling how far a score is from the mean in standard
deviation units. The formula to convert any score (x) into its corresponding z score is
( x − x)
z= ,
s
where x is the observed score and s is the standard deviation of the distribution. If for
example, the following observations are given, then z scores can be calculated for
selected observations (Table 3.3).
s =
∑ (x - x ) 2
n -1
29.5
=
6 -1
= 2.42
Here, for x = 8, what will be the corresponding z score? Using the above data,
( x − x)
z =
s
(8 − 11.5)
=
2.42
= -1.45
37
Biometry
In the same way, for x = 15, the z score will be z = 1.45. When a z score is positive, it
is located above the mean, and when negative it is below the mean. if the population
parameter µ and σ are known, the z score can also be calculated as:
(x − µ)
z= .
σ
Therefore, the new variety did best at Melkasa and poorest at Bako.
38
Biometry
There are basic questions that the answers will help to decide whether an analysis is
worthwhile or not. First, what are the goals of analysis? There are times when very
little is known about the variables and relationships between variables. Secondly, a
vague goal might be stated, such as learning about the variables and about
relationships between these variables. The vague idea can be developed into new
goal being revamped, which in turn can provide the fundamental thinking behind the
experiment. Thirdly, the vague idea becomes rephrased as more formal questions.
For example, the goal of a research program is to increase yield of rice. Here, the
question is narrowed down to how is the crop best managed to obtain the best yield,
more specifically what types and rates of fertilizer can give the best yield of rice at a
given location? Then the objective will be to determine the level of nitrogen fertilizer
that provides maximum yield of a crop on three improved varieties, for example.
The next question is: can the data set be used to meet the goal of the analysis? This
is a question often overlooked by researchers and data analysts. There are cases
when the variables recorded cannot legitimately meet the goals of the research or the
data are so poorly collected that the analysis is worthwhile. We have to be careful on
the collection of data so that it can be useful in analysis, interpretation and writing
results. Recognize the limitation of our data; we can set limits on our interpretations
of analysis.
The following question is: how can a data set are obtained for analysis? Data must
be collected, compiled and organized before going for analysis. Each variable should
be measured and recorded correctly. The collected data should be formatted the way
the programme for analysis requires. Mistakes in inputting data from data forms to a
computer file must be avoided. After proof reading, error checking and formatting
analysis of variance can be done. What is analysis of variance (ANOVA)?
The ANOVA is a method that can be used to analyze data that resulted from both
simple and complex experiments (Cochran and Cox, 1957). It is one of the most
39
Biometry
important statistical techniques available and provides a link between the design of
experiments and the analysis of experimental data. The ANOVA has its origin in
biology, or at least agriculture, since the methods were specifically developed to deal
with the type of variable response that is common in field experiments but currently it
has a much wider application (Steel and Torrie, 1980).
In a single factor experiment, there are several independent random samples that
can be used to make inferences about the populations from where the samples are
extracted. Here, the concern is in making inferences about means of the populations
sampled.
The objective of this Section is to enable the reader explain the purpose of ANOVA,
identify the assumptions that underline the ANOVA technique, describe the ANOVA
hypothesis-testing procedure and use the ANOVA testing procedure to arrive at
statistical decision about the means of treatments.
There are a number of advantages with the CRD. It is flexible in terms of the number
of treatments in each replication, simple for statistical analysis, provides maximum
degrees of freedom for error and missing data has no effect on the analysis. The
primary disadvantage is that it gives less precise results in heterogeneous
experimental units. Another disadvantage is that all variations other than that of the
treatment are considered as experimental error.
4.2.1. Layout
There is no restriction on the assignment of treatments to experimental units (plots) in
CRD. The only situation is that treatments should be assigned to plots at random. For
example, for t number of treatments to be assigned in n number of plots where, n > t,
there will be tr number of plots, if each treatment is equally repeated (r). If there are
different repeats for different treatments, then n = ∑ti ri.
40
Biometry
1 2 3 4
5 6 7 8
9 10 11 12
Figure 4.1. Field plot plan for 12 plots in CRD
Then, randomization is applied using the random number table as follows. First, a
starting point is selected anywhere in a table of random numbers (Appendix 1).
Secondly, the first 12 digit numbers are written by moving up or down, left or right.
Here, it is better to use three digit numbers for the reason that they are less likely to
contain the same number than when numbers with fewer digits are used. Thirdly, the
numbers are ranked from smallest to largest as in third column of Table 4.3. These
ranks correspond to plot numbers on the field plot plan. Fourthly, the first treatment is
assigned to the first four plots in sequence, the second treatment to the second four
plots and the third treatment to the remaining plots.
The resulting randomization would look like Table 4.3 and the master field plan would
be as in Figure 4.2.
41
Biometry
4.2.2. Analysis
The analysis can be started by calculating the total amount of variability in the data,
irrespective of the treatments. The best way to do this is to calculate a sum of
squares (SS) as a measure of variability. That is the total sum of squares (Total SS).
Recalling that a sum of squares (SS) is calculated from ∑(x - x )2, if there is a need to
calculate the Total SS, the mean to be used in this case will be the grand mean of all
of the data. Detailed calculations are shown later. The total variability, measured by
the Total SS, can be partitioned into two components: the explained and the
unexplained variability. The explained variability is that represented by differences
between the treatments. This variability is explained because there is a potential
explanation for it, i.e. the treatments applied to the experimental units were not the
same. The unexplained variability is that within the treatments, i.e. differences
between replicates. This variability is not caused by the experimental treatments.
Hence, it is often called experimental error. Thus, the following important relationship
holds:
Since the values of two elements in the summation are known, the third value can be
obtained by subtraction. If Total SS = Explained SS + Unexplained SS then
unexplained SS = Total SS - explained SS.
F = Explained MS/Unexplained MS
The terms Treatment and Error are often given other names depending on the
authors’ notation. So it is advisable to provide readers with different terms to mean
the same concept. Treatment is synonymous with factor or between and error is
synonymous with residual or within. There are few basic computations to be done.
42
Biometry
These are estimating the treatment effects, standard error of the mean or difference
and test for significant differences among the treatments. These can be done
following the procedure of the ANOVA. In this procedure, the components should be
explained by linear model at the beginning.
The linear additive model (L.A.M.) for CRD is Xij = µ + Ti + Eij where, Xij = the jth
observation of ith treatment, µ = the overall mean, Ti = ith treatment effect (µi-µ) and Eij
= the effect of jth observation of ith treatment. j=1…r, I=1…t.
The following assumption must be made concerning the random error in order to use
the one-way ANOVA technique:
a. Each treatment population has the same variance, that is, the random errors
have the same variance in experiments designed to compare treatment
means,
b. The error terms are statistically independent, that is, the magnitude of the error
in one observation is not influenced by the magnitude of the error in an other
observation,
c. The errors are normally distributed. These assumptions can be abbreviated as
follows and the abbreviations will be used here after.
Assumptions
Ti ~ NI (0, σ2t)
Eij ~NIR (0, σ2)
Practical situations occur where one or more basic assumptions of the ANOVA are
violated. For example, the assumption of independence is not true when the errors
are correlated. The most common causes of positive correlation between errors are
time and space trends in experimental materials, environmental variables or tools
(Steel and Torrie, 1980). Negative correlation may also occur in the case of
compensating errors and is seen in data such as those taken for successive batch
yields from multistage process (Cochran and Cox, 1957). The elimination of
correlated errors is best achieved by removing the sources of correlation. This can be
achieved through randomization in sampling to average out the trends and allow us
the ANOVA technique (Sokal and Rohlf, 1987).
In many cases the equal-variance assumption does not hold as a result of correlation
between the mean and the standard deviation. This relationship can be eliminated
and the variance made uniform by a simple data transformation (Cochran and Cox,
1957). Correlation between the mean and standard deviation is frequently found in
samples exhibiting non-normality. If the errors are not normally distributed the true
probability of accepting the hypothesis that there is no treatment effect, is not equal
to the probability given by significant tables.
Most statisticians agree that one or more assumptions are violated to some degree
(Cochran and Cox, 1957; Little and Hills, 1978; Steel and Torrie, 1980; Peterson,
1985 and Sokal and Rohlf, 1987). The violation in a given research setting affects the
validity of inferences under some statistical procedures more severely than others.
43
Biometry
The primary interest here is the validity of inferences under the ANOVA, because the
assumptions are the basic tool of analyzing data from designed experiments. Thus,
the concern is whether or not the underlying assumptions of the ANOVA
methodology are violated to the extent of appreciably changing the error rates.
Several authors reported on the validity of inferences from the ANOVA and their
report can be summarized as follows:
a. The classical ANOVA F-test is very robust with respect to normality.
b. When there are equal replications, mild departure from homogeneous
variance has little impact on the validity of the ANOVA test.
c. Unequal variances have a greater impact on estimates of treatment effects
and their variances than on tests.
Replications
Treatment 1 2 3 …… r Treatment Total (Ti)
1 X11 X12 X13 …… X1r T1
2 X21 X22 X23 …… X2r T2
3 X31 X32 X33 …… X3r T3
. . . . . .
. . . . . .
. . . . . .
t Xt1 Xt2 Xt3 …… Xtr Tt
Grand Total GT
The next step is to set up analysis of variance table as in Table 4.5a to 4.5d. In the
ANOVA tables, the total variation is partitioned into two components: variation due to
treatment and variation within treatment (experimental error/residual). The degree of
freedom for each component is one less than the number of observations. That is t-1
for treatment, tr-1 for total and t(r-1) for error with equal repeats or n-t for unequal
number of repeats.
44
Biometry
The computations are easier if the treatments are equally replicated. Both conditions
will be explained by taking treatments with equal and unequal replications, for the
reason that CRD with different number of replications is not commonly found in text
books.
Table 4.5a. Analysis of variance for CRD where treatments have different replication
Table 4.5b. Analysis of variance for CRD where treatments have equal replications
Table 4.5c. Expected mean square for CRD of equal replications (Random model)
Source of Expected
variation d.f. Mean Squares E(MS)
Treatments t-1 σ2e + rσ2t
Error t(r-1) σ2e
Table 4.5d. Expected mean square for CRD of equal replications (Fixed model)
Source of Expected
variation d.f. Mean Squares E(MS)
Treatments t-1 σ2e + r∑(xi- x ) 2/t-1
Error t(r-1) σ2e
45
Biometry
After the ANOVA is done, F-calculated is tested against the tabulated F-value at the
desired level of probability. F = MSt/MSE. F-values, unlike t-values, have 2 parts to
the degrees of freedom: n1 = treatment d.f.; n2 = error d.f. F-tables are arranged with
the treatment degrees of freedom horizontally and the error degrees of freedom
vertically as in Appendix 4a and 4b.
If the F-calculated is greater than F-tabulated, then it will be concluded that there is a
significant treatment effect. Comparison between calculated and tabulated F-vales
will lead to either rejecting or accepting the null hypothesis. The hypothesis can be
framed as H0: all treatments have equal means and the alternate hypothesis as H1: at
least one treatment mean is different from other treatment means. On the other hand
the MSE = s2 is the sample estimate of experimental error that can be used to
compute standard errors and interval estimates. The key issue here is that standard
errors are computed and intervals will be estimated, if the null hypothesis (H0,) is
rejected, that is, the Mean Squares of treatment are found to be significant.
The purpose of the second category of mean separation is to test mean difference
between any treatment and a specified treatment usually a control or standard check.
In this case, there are only t-1 paired comparisons (Peterson, 1977). This category
includes preplanned contrast and preplanned comparison to a check or control.
46
Biometry
The least significant difference (LSD) procedure, also known as Fisher’s LSD,
compares all possible [t(t-1)/n] pairs of standard t-test for t treatment means. The
LSD procedure is only used when the treatment source of variation is found to be
significant by the F-statistic. The purpose of testing each mean difference is to test
the null hypothesis that the corresponding two treatment means are equal.
The critical value for conducting each test is the same for each individual comparison
of paired means, and it depends on the number of degrees of freedom for the error,
as well as the significance level of the test, α. It should be noted that α reflects the
probability of a Type-I error for each individual comparison, and therefore cannot be
applied to the entire group of individual comparison since these are not mutually
independent. Thus, the t-tests of all possible hypotheses are not independent.
The LSD method differs from others in that it has a comparison-wise Type-I error, α,
over all repetitions of the experiments (Steel and Torrie, 1980). This means that the α
risk will be inflated when using the LSD procedure. As the number of treatments
increases, the Type-I error for experiment becomes large. Due this fact, a situation
may arise whereby the F-statistic in the ANOVA is significant, yet the LSD procedure
fails to find any pair of treatment means which differ significantly from one another.
This occurs since the F-statistic is considering all possible comparisons between
treatment means simultaneously, not in a pair-wise manner as the LSD procedure
does (Cochran and Cox, 1957).
2MS E
LSD = t(1- α)/2
r
Assuming two-sided alternatives, the pair of means would be declared significantly
different if the difference between the two means is greater than the LSD value.
2MS E
LSD = (tα/2, error d.f.) for equal replication and
r
1 1
LSD = (tα/2, error d.f.) 2MS E ( + ) for unequal replication.
ri rj
Taking five treatments, for example, the process can be explained as T1, T2, T3, T4
and T5 with mean values of 10, 11, 15, 17 and 22, respectively, with α= 0.05, error
degrees of freedom 20, and MSE for the five treatments = 8.06, then the critical
difference = 4.37. From this statistics one can see that treatment 5 had the highest
mean and two pairs of means that do not significantly differ from each other are T1
and T2 and T3 and T4.
T1 T2 T3 T4 T5
9 10 16 17 22
The overall α risk may be considerably inflated using this method. Specifically, as t
getting larger, the type I error of the experiment becomes larger (Steel and Torrie,
1980).
47
Biometry
b. Tukey’s Test
Tukey (1953) proposed a procedure for testing hypotheses for which the overall
significance level is exactly α when the sample sizes are equal and at most α when
the sample sizes are unequal. His procedure can also be used to contract confidence
intervals on the differences in all pairs of means. For these intervals, the
simultaneous confidence level is 100(1- α) percent when the sample sizes are equal
and at least 100(1- α) percent when the sample sizes are unequal. This is good
procedure when interest focuses on pairs of means. Tukey’s procedure makes use of
the distribution of studentized range statistic
Appendix Table A.8 of Steel and Torrie (1980) contains values of qα (p, d.f.), the
upper α and percentage points of q where d.f. is degrees of freedom for error. For
equal sample size, Tukey’s test declares two means significantly different from each
other, if the absolute value of their sample differences exceeds Tα.
MS E
Tα = qα(t, d.f.) .
r
In the same way, we can construct a set of 100(1- α) percent confidence intervals for
all pairs of means. For the previous example, with means 9, 10, 16, 17 and 22, with
α= 0.05 and error degrees of freedom 20, there will be q0.05 (5, 20) = 4.23 and if the
MSE for the five treatments = 8.06, then
8.06
qα (5, 20) = 5.37.
5
Hence, any pair of treatment means that differ in absolute value by more than 5.37
would imply that the corresponding pair of population means is significantly different.
T1 T2 T3 T4 T5
9 10 16 17 22
48
Biometry
MS E
SE(m) = .
r
From Duncan’s Table of significant ranges, Appendix Table A.7 of Steel and Torrie
(1980), rα (p, d.f.) values are obtained for t = 2, 3, …, t, where α is the significant level
and d.f. is error degrees of freedom. This ranges will be converted into a set of t-1
least significant ranges (Rp) for p = 2, 3, …, t by calculating Rp).
Then, the observed differences between means are tested, beginning with the largest
versus smallest, which would be compared with the least significant ranges Rt. Next,
the difference of the largest and second smallest is computed and compared with the
least significant range Rt-1. Such comparisons are continued until all means have
been compared with the largest mean. Finally, the difference between the second
largest mean and the smallest is compared against the least significant range Rt-1.
This process is continued until the difference between all possible n(n-1)/2 pairs of
means have been considered. If the observed difference is greater than the
corresponding least significant range, we conclude that the pair of means in question
is significantly different. To prevent contradictions, no differences between a pair of
means are considered significant if the two means involved fall between two other
means that do not differ significantly.
DNMRT can be applied on the previous example. Recalling that MSE = 8.06, n = 5
and error degrees of freedom = 20, then the treatment means can be ordered in
ascending order as T1=9, T2 = 10, T3 = 16, T4 = 17 and T5 = 22.
MS E
The SE (m) =
r
8.06
=
5
= 1.27.
From the table of significant ranges in Appendix Table A.7 (Steel and Torrie, 1980)
for 20 degrees of freedom and 0.05 level of probability, we obtain r0.05(2, 20) = 2.95,
r0.05(3, 20) = 3.10, r0.05(4, 20) = 3.18 and r0.05(5, 20) = 3.25. Hence, the least
significant ranges are
R2 = r0.05(2, 20)SE(m) = 3.75
R3 = r0.05(3, 20) SE(m) = 3.94
R4 =r0.05(4, 20) SE(m) = 4.04
R5 = r0.05(5, 20) SE(m) = 4.13
The comparison would produce that there are significant differences between all
pairs of means except treatment 1 and 2, and 3 and 4. In this example, DNMRT and
49
Biometry
the LSD method produced the same result that leads to identical conclusions. DMRT
requires a greater observed differences to detect significantly different pairs of means
included in the group. For two means the critical value of DNMRT and the LSD are
exactly equal. Generally, DNMRT is quite powerful when n is getting larger; that is, it
is very effective at detecting differences between means when real differences exist
and for this reason it is popular.
Some statisticians use Tukey’s method as it does control the overall error rate. The
Newman-Keuls Test is more conservative and the power of the test is less than
DNMRT. There are other multiple comparison procedures and further references can
be obtained in Steel and Torrie (1980) and other texts of experimental designs.
50
Biometry
chosen manufacturers (treatments). The data in hours to first failure is given below
(Table 9) for the three models (manufacturers’ printers).
Model-I 60 45 72 68 71 52 368
Model-II 102 96 105 99 103 95 600
Model-III 121 132 118 128 131 126 756
Grand Total 1724
In this example, the shortcut methods will be introduced as they are of practical
utilities. The stepwise computations of the different components are as follows.
SSt =
∑ Ti 2 – C.F.
r
[(368) 2 + (600) 2 + (756) 2 ]
= - 165120.89
6
=12705.78
5. Sum squares of error (SSE) = Total SS - SSt
= 13547.11 - 12705.78
= 841.33
Table 4.10. Analysis of variance for CRD with three treatments and of six repeats
Source of
51
Biometry
The formula to be used will be the one with equal number of repeats. A common
standard error of difference will be calculated to serve all treatment means, but in the
previous case where there were different repeats, a standard error of difference for
each treatment means was computed.
52
Biometry
The conclusion from this mean comparison is that all models are statistically different
where the highest mean first failure value was due to Model-III. Hence, the company
can go for Model-I system option to develop a laser quality printer as it has lowest
mean value of first failure.
The principal advantages of RCBD are: it is more precise than CRD when block
effect is significant, any number of treatments and replications can be included, the
statistical analysis is easy, and it provides information on the uniformity of
experimental units. This design is one of the most frequently used designs as it
provides a required precision with limited cost.
In this design, if each treatment is applied to exactly one unit in the block and
comparisons are only drawn between treatment responses from the same block,
extraneous variability should be reduced. This is the concept underlying the
randomized complete block design.
A few typical examples for which the RCBD may be appropriate are performance
evaluation of newly developed varieties against the standard check; a specific
fertilized rate determination trial on a specific crop; animal nutrition experiment to
identify the best ration (using body weight or age as blocks); clinical trials to compare
several competing drugs (using age or severity of symptom as blocks); psychological
experiments comparing several stimuli (using socioeconomic background as a block);
comparison of several techniques for storing fruits or vegetables (using each
incoming shipment as a block); and so on.
4.3.2. Layout
Randomization is a basic part of the RCBD. After the experimental area is grouped
into blocks and plots, the treatments are assigned at random to the plots within each
53
Biometry
block. The same procedure is repeated with a new randomization for each of the
remaining blocks.
Suppose there are six treatments to be evaluated in RCBD with four replications
under field condition. Then, the field plan can be constructed as follows. Assuming
the gradient as soil fertility that runs from top to bottom, blocking is done
perpendicular to the gradient, in this case left to right as shown in Figure 4.3 and the
master plan can be done as in Table 4.17, respectively.
Soil fertility
High Block-I
Plot 101 102 103 104 105 106
Block-II
Plot 201 202 203 204 205 206
Block-III
Plot 301 302 303 304 305 306
Block-IV
Plot 401 402 403 404 405 406
Low
Figure 4.3: Field plot plan for evaluating six treatments in RCBD
4.3.3. Analysis
Once the data are obtained, they can be arranged in a two-way table, where rows
represent treatments and columns represent blocks or vice versa. Depending on the
size of the block used, there are two basic types of randomized block designs:
complete block design and incomplete block design. Complete block design refers to
a situation where each block contains all the treatments while in the case of
incomplete block designs at least the size of one block is less than the number of
treatments in the experiment. As the incomplete block designs are illustrated in
Section 4.6, the discussion is limited to RCBD in this Section.
If the measurements that corresponds to treatment i and block j are designated by xij,
the data structure of the RCBD with b block and t treatment is shown in Table 4.18.
The arrangement of the data depends on personal preference. The column can be
54
Biometry
assigned for the treatments and the rows for the blocks. If the numbers of treatments
are more than the number of replication it is better to use the rows for treatment and
column for blocks.
Table 4.19a. ANOVA Table for a RCBD with b blocks and t treatments
Source of
variation d.f. SS MS F-cal
Blocks b-1 SSB MSB MSB/MSE
Treatments t-1 SSt MSt MSt/MSE
Error (b-1)(t-1) SSE MSE
55
Biometry
Table 4.19b. Expected mean square for a RCBD with b blocks and t treatments
Source of
variation d.f. E(MS) Random E(MS) Fixed
Blocks b-1 σ2 + tσR2 σ2 + t(Rj2)/(r-1)
Treatments t-1 σ2 + rσT2 σ2 + r(Ti2)/(t-1)
Error (b-1)(t-1) σ2 σ2
Table 4.23. Data from six populations of maize evaluated for 100-grain weight in
Serdang, Selangor.
1. The linear additive model (L.A.M.): Xij = µ + Ti + Bj + Eij where, Xij = the ith
treatment effect in jth block, µ = the overall mean, Ti = ith treatment effect (µi-µ), Bj
is jth block effect (µj-µ) and Eij = the effect of ith treatment in jth block. j=1…r,
i=1…t.
Assumptions
Ti ~ NI (0, σ2t)
56
Biometry
Bj ~ NI (0, σ2B)
Eij ~ NI (0, σ2)
(GT ) 2
2. Correction factor (C.F.) = where, r= number of replication and
rt
(644.1) 2
= t = number of treatment
4 x6
=17286.03375
3. Total sum of squares (Total SS)
Total SS = ∑∑Xij 2 –C.F.
=(27.0)2 + (30.4)2 + … + (23.3)2 – 17286.03375
=225.856
4. Sum of squares due to treatment (SSt)
2
SSt =
∑ Ti
– C.F.
r
[(114.8) 2 + (118.1) 2 + ... + (87.9) 2 ]
= - 17286.03375
4
=155.68375
5. Sum of squares due to block (SSB)
2
SSB =
∑ Bj
– C.F.
t
[(159) 2 + (166.8) 2 + ... + (163.8) 2 ]
= - 17286.03375
6
= 14.62125
6. Sum of squares due to error (SSE) = Total SS-SSt-SSB
= 225.856-155.68375-14.62125
= 55.55129
7. Degrees of freedom for treatment (d.f. for t)
d.f. for t = number of treatment (t)-1
= 6 -1
=5
8. Degrees of freedom for block (d.f. for b)
d.f. for b =b-1
=4-1
=3
9. Degrees of freedom for error (d.f. for e)
d.f. for e = (t-1)(r-1)
= (6-1)(4-1)
= 15
10. Mean square due to treatment (MSt)
MSt = SSt/d.f. for t
= 155.68375/5
= 31.13675
11. Mean square due to block (MSB)
MSB = SSB/d.f. for b
57
Biometry
= 14.62125/3
= 4.876
12. Mean square due to error (MSE)
MSE = SSE/d.f. for e
= 55.55129/15
= 3.7034
13. Calculate F-values
a. Replication
F-cal = MSB/MSE
= 4.87375/3.7034
= 1.316
b. Treatment
F-cal = MSt/MSE
= 31.13675/3.7034
= 8.4076
Then the calculated F-vales are compared with the tabulated F-values at 0.05 and
0.01 level of probability: Replication with d.f. 3 and 15 = 3.29 and 5.42 and for
treatment with d.f. 5 and 15 = 2.90 and 4.56, respectively. For replication, the
calculated F-value is less than the tabulated F-value while for treatment; the
calculated F-value is greater than the tabulated F-values (Table 4.24). The tabular F-
values are obtained using the degrees of freedom for replication or treatment as V1
value and error degrees of freedom as V2 value. For example, the tabulated value V1
= 5 (degrees of freedom for treatment) and V2 = 15 (degrees of freedom for error) are
2.90 at 0.05 and 4.56 at 0.01 level of significance.
Source of F-tabulated
variation d.f. SS MS F-calc. 0.05 0.01
Block 3 14.62125 4.876 1.32 3.29 5.42
Population 5 155.68375 31.13675 8.41** 2.90 4.56
Error 15 70.17254 3.7034
58
Biometry
Once the CD value is obtained, paired population means can be compared after
computing the number of possible pairs of means. Number of possible combination =
n (n-1)/2 = 6(6-1)/2, where n is the number of treatments (Table 4.25). The pair wise
comparison below can be summarized as
T2a T2
T1a T1
T3a or T3
T5ab T5
T4 b T4
T6 c T6
Treatments connected with the same bar or having the same letter show non-
significant difference among them.
59
Biometry
When the degrees of freedom for error is less than 20 Gomez and Gomez (1984)
suggested that the R.E. value should be multiplied by a correction factor (K) defined
as
[(r − 1)(t − 1) + 1][t (r − 1) + 3]
K=
[(r − 1)(t − 1) + 3][t (r − 1) + 1]
[(4 − 1)(6 − 1) + 1][6(4 − 1) + 3]
=
[(4 − 1)(6 − 1) + 3][6(4 − 1) + 1]
= 0.9824
The value 1 is subtracted and then the efficiency is expressed as percentage, i.e., it
will be 1-1.0228 = 0.0228. When 100 multiply it, it will be 2.3% indicating that the use
of RCBD instead of CRD increased experimental precision by 2.3% which is quite
low.
60
Biometry
Interpretation
The analysis of variance for RCBD (Table 4.24) showed that the effect due to the
populations of maize was highly significant (p=0.01). Synthetic GxA had the highest
mean 100-grain weight (though Synthetic GxA, Suwan-I, Hercules and Venus 49
were not statistically different) followed by Suwan-I and Hercules (Tables 4.25a and
4.25b). Synthetic GxA, Suwan-I and Hercules were all better than Red Iron 45 and
MARDI composite-I. Therefore, Synthetic GxA, Suwan-I and Hercules are potential
populations for further evaluation.
Blocking effect was not significant indicating that blocking was not necessary in
controlling the gradient considered during planning or the gradient was less
heterogeneous. The relative efficiency of RCBD was very small (2.3%) showing that
RCBD and CRD had almost similar precision.
61
Biometry
4.4.2. Layout
The basic pattern of Latin square design can be shown using a 4 x 4 Latin square. In
this case the number of treatments, t equals 4, denoted as A, B, C and D, as shown
in Figure 4.4. From the Figure it’s clear that each row and column form a complete
block as in a randomized complete block design. The two modes of blocking help to
reduce sources of variations and thereby improve precision. Randomization can be
done in such a way that the first row is arranged in alphabetical order followed by
shifting subsequent letters one position to the left as ABCD, BCDA, CDAB and
DABC. The order of the rows can also be arranged as 2, 1, 3 and 4, followed by
randomization of the order of columns as 4, 2, 3 and 1 to get the following
randomization. Finally, treatments are assigned to plots.
Column
Row 1 2 3 4
1 A C D B
2 D B C A
3 B D A C
4 C A B D
Figure 4.4. Basic layout of a 4 x 4 Latin square
4.4.3. Analysis
The analysis of data from Latin square follows much the same pattern as for RCBD
except that one source of variation is added. For example, for t treatments that will
occupy t2 experimental units. Then the data collected can be organized as in the
Tables 4.41 and 4.42.
Table 4.41. Data structure for Latin square for column C and row R
Row Column Rj
1 2 . . . k
1 x11 x12 . . . x1b R1
2 x21 x22 . . . x2b R2
3 x31 x32 . . . x3b R3
62
Biometry
. . . . . . . .
. . . . . . . .
. . . . . . . .
j x1j x2j . . . xkj Rt
Column total C1 C2 . . . Ck GT
Table 4.42 Data structure of Latin square for treatment totals (Ti)
Treatment 1 2 . . . t
Sum T1 T2 . . . Ti
Ti is obtained by adding all values in each column/row where that particular t appeared
63
Biometry
c. Treatment = MSt/MSE
Then the calculated F-vales are compared with F-tabulated at 0.05 and 0.01 level of
probability. All components of the ANOVA and expected mean squares are shown in
Table 4.43, 4.44 and 4.45.
Source of
variation d.f. SS MS F-cal
Rows r-1 SSR MSR MSR/MSE
Columns c-1 SSC MSC MSC/MSE
Treatments t-1 SSt MSt MSt/MSE
Error (t-1)(t-2) SSE MSE
For all the three components the degrees of freedom is the same in Latin square and can be given by
simply t-1 for each of them, i.e., row, column and treatment have the same degrees of freedom.
Source of
variation d.f. E(MS) -Random E(MS) -Fixed
Rows r-1 σ2 + tσR2 σ2 + t((∑Rj 2)/(t-1)
Columns c-1 σ2 + tσC2 σ2 + t(∑Ck 2)/(t-1)
Treatments t-1 σ2 + tσT2 σ2 + t(∑Ti 2)/(t-1)
Error (t-1)(t-2) σ2 σ2
Table 4.45. Expected mean squares for Latin square design with squares/sets
Source of
variation d.f. E(MS) Random
Squares/sets s-1 σ2+ tσS2
Rows/sets (r-1)s σ2 + tσ2R/S
Columns/sets (c-1)s σ2 + tσ2C?S
Treatments t-1 σ2 + tσT2
Error s(t-1)(t-2) + (s-1)(t-1) σ2
64
Biometry
MS E
11. Coefficient of variation (c.v.%) = 100( ).
GM
[ MS R + MS E (t − 1)]
Over RCBD, RE(RCBD, row) =
t ( MS E )
[ MSC + MS E (t − 1)]
Over RCBD, RE(RCBD, column) =
t ( MS E )
Example-I
There are experimental animals that do differ in body weight as well as in age (two
gradients). A researcher, planning to compare 5 newly formulated rations and a
check, designed his experiment in Latin square design. Body weight was labeled as
row and animal age as column. After a certain period of time, the weight gain was
recorded as in the following Table 4.46.
Column
Row 1 2 3 4 5 6 Rj Mean
1 F(2.19) E(2.50) D(2.27) C(1.62) B(1.82) A(0.91) 11.31 1.89
2 E(2.27) C(1.41) A(0.91) D(1.91) F(2.13) B(1.95) 10.58 1.76
3 B(2.04) A(0.91) F(2.25) E(2.29) D(1.91) C(2.07) 12.06 2.01
4 A(0.77) (B2.04) E(2.40) F(1.99) C(1.82) D(2.50) 11.52 1.92
5 D(2.50) F(2.31) C(2.09) B(2.04) A(0.91) E(2.27) 12.12 2.02
6 C(1.52) D(1.86) B(1.91) A(0.77) E(2.30) F(1.98) 10.34 1.72
Ck 11.29 11.03 00.83 10.62 11.48 11.68 67.93 1.89
65
Biometry
(GT ) 2
1. Correction factor (C.F.) = , where, t = number of treatments
t2
(67.93) 2
=
62
=128.18
2. Total sum of squares (Total SS)
Total SS = ∑∑Xijk 2 –C.F.
= (2.19)2 + (2.27)2 + … + (1.98)2 – 128.18
= 9.93
3. Sum of squares due to Row (SSR)
2
SSR =
∑ Rj
– C.F.
t
[(11.31) 2 + (10.58) 2 + ... + (10.34) 2 ]
= – 128.18
6
= 0.46
4. Sum of squares due to Column (SSC)
2
SSC =
∑ Ck
– C.F.
t
[(11.29) 2 + (11.03) 2 + ... + (11.68) 2 ]
= – 128.18
6
= 0.17
5. Sum of squares due treatment (SSt)
2
SSt =
∑ Tk
– C.F.
t
[(5.18) 2 + (11.80) 2 + ... + (12.85) 2 ]
= – 128.18
6
= 8.86
The treatment total is obtained by adding all observations for each treatment as
shown below:
• Total of treatment A = 0.77+0.91+0.91+0.77+0.91+0.91 = 5.18
• Total of treatment B = 2.04+2.04+1.91+2.04+1.82+1.95 = 11.80
• Total of treatment C = 1.52+1.41+2.09+1.62+1.82+2.07 = 10.53
• Total of treatment D = 2.50+1.86+2.27+1.91+2.50+2.50 = 13.54
• Total of treatment E = 2.27+2.50+2.40+2.29+2.30+2.27 = 14.03
• Total of treatment F = 2.19+2.31+2.25+1.99+2.13+1.98 = 12.85
66
Biometry
Then the calculated F-vales are compared with tabulated F-value at 0.05 and 0.01
level of probability: For row and treatment the calculated F-value is greater than the
tabulated F-value (2.71 at 5%) while for column the calculated F-value is less than
the tabulated F-value (Table 4.47).
Source of F-tabulated
variation d.f. SS MS F-cal 0.05 0.01
Rows 5 0.46 0.092 4.18** 2.71 4.10
Columns 5 0.017 0.034 1.55 2.71 4.10
Treatments 5 8.86 1.772 80.55* * 2.71 4.10
Error 20 0.44 0.022
MS E 0.022
14. Coefficient of variation (c.v.%) = 100 ( ) = 100 x (
) = 7.8%.
GM 1.89
Row and treatment effects are significant and hence, we proceed to calculate
standard error of the mean, standard error of difference and critical (least
significant) difference to identify the best ration.
67
Biometry
[ MS R + MSC + MS E (t − 1)]
Over CRD, RE(CRD) =
[(t − 1) MS E
[0.092 + 0.034 + 0.022(6 − 1)]
=
[(6 − 1)0.022
= 1.53 or 53%
[ MS R + MS E (t − 1)]
Over RCBD, RE(RCBD, row) =
t ( MS E )
[0.092 + 0.022(6 − 1)]
=
6(0.022)
= 1.53 or 53%
[ MSC + MS E (t − 1)]
Over RCBD, RE(RCBD, column) =
t ( MS E )
[0.034 + 0.022(6 − 1)]
= 6(0.022)
= 1.09 or 9%
The use of LSD increased precision by 53% over RCBD, but the addition of column
blocking in the LSD did not increase precision. Hence, RCBD with rows as blocks
would have been as efficient as LSD. Since there was a significant difference among
treatments, it is necessary to do mean separation to identify the best treatment
(Table 4.48) and then present the data as in Table 4.49.
68
Biometry
Table 4.49. Mean body weight gain (kg) due to different rations
Ration Mean
A 0.86
B 1.76
C 1.97
D 2.26
E 2.34
F 2.14
5% CD 0.19
CV 7.8%
Interpretation
The treatment effect was significant indicating that there is a significant difference
among the treatment means (Table 4.47). Those treatments that had their mean
difference greater than the critical value are statistically different where as those
having their difference less than the critical values were not statistically different.
All the new formulations were superior to the standard ration (treatment A).
Treatment E gave the highest (although treatment E and D are not statistically
different) mean body weight followed by treatment D and F. Treatment E had 172%
advantage in increasing the body weight of animals compared to the standard ration,
treatment A. Hence, treatment E is a potential ration for use (Table 4.49).
The effect due to rows was significant implying that blocking in the row direction
significantly reduced the variation considered (initial body weight). However, the
effect due to column was not significant indicating that considering age as a source
of variation was not important in this particular experiment. Therefore, randomized
complete block design with blocking in the row direction (blocking the animals by
initial body weight) could have been satisfactory.
69
Biometry
In the early stages of selection process there could be insufficient seed of the new
progenies (genotypes) to undertake replicated experiments or the number of
genotypes could be very large to manage in terms of resources. In such cases, some
plant breeders use single row plots to evaluate the newly developed progenies and at
a certain regular intervals check varieties are planted. The performances of the new
genotypes are then compared with the performance of the nearest check,
subjectively. The standard statistical analysis is not possible as the new genotypes
are not replicated but an objective comparison can be made. The disadvantage to
the subjective judgment is that the checks are systematically placed and no provision
is made to adjust a given measurement for environmental variation from one part of
the experiment to another.
A better method, when there are many progenies at the early stage of the breeding
programme, is to use an experimental design called augmented design that was
developed by Federer and well illustrated by Federer and Ragavarao (1975). This
design is of particular interest in an extensive plant breeding programme.
4.5.2. Layout
The first step is to divide the experimental area into blocks and then assigning plots
within the blocks in order to have a single row representing a treatment. Two or more
checks are assigned to each block at random and the remaining rows are assigned
to the new genotypes. The design is more efficient if blocks of the same size are
used but it does not necessarily be the same size. The resulting design is one in
which the checks are replicated but the new genotypes are not. By doing so the
design helps to estimate experimental error and provides a means to adjust the yield
of the new genotypes for variations from block to block.
The number of checks and blocks is determined by the statistically acceptable error
degrees freedom, which are at least 12. This restriction determines the minimum
number of blocks and checks to be used. The maximum number of blocks depends
on the number of new genotypes and the experimental area. If for example, there are
c number of checks and b number of blocks, then the degrees of freedom for error is
(c-1) (b-1). Hence, the number of blocks can be equated as follows: b ≥ 12/(c-1) + 1.
The minimum number of blocks for 3 checks will be 7, for example.
4.5.3.Analysis
In the data analysis, the experimental error is estimated considering the checks
placed in randomized complete block design, which has been dealt in detail in
Section 4.3. After conducting the analysis on the checks, the error mean square from
70
Biometry
this analysis is used to compute the standard errors for different comparisons. The
difference between the mean of the checks in a given block and the mean of the
checks over the entire experiment.
The first step in the analysis is to conduct ANOVA on the data measured for checks.
Then, a table is constructed (Table 4.61) to organize the different entities (check,
block, sum for the checks & blocks, mean for the checks & blocks and the
adjustment). Let Xij be the yield of the ith check in jth block. Then, adjustment aj =
mean of j – the grand mean and the sum of aj = 0. The variable of interest (example
yield) of the new genotypes is adjusted for the effect of the block in which it was
grown as shown in Table 4.62.
71
Biometry
SSR =
∑ R j 2 – C.F.,
c
2
SSC =
∑ Ci
– C.F.,
b
SSE = Total SS – SSR – SSC and
SS E
Mean square of error (MSE) = .
(c − 1)(b − 1)
The analysis of variance for the checks can be given as in Table 4.63
Source d.f. SS MS
of variation
Blocks b-1 SSB
Checks c-1 SSC
Error (b-1)(c-1) SSE MSE
Standard errors
There are four types of standard error of differences to compare the means in
augmented design. These standard error of differences are computed as follows. To
compare two check means standard error of difference:
2 MS E
SE(d) = .
b
72
Biometry
These standard errors of differences can be used to compute the least significant
difference (LSD) values.
4.5.4.Numerical example
P1 A P2 B P3 C P4 D Total
120 83 100 77 90 70 85 65 690
P5 B P6 C P7 A P8 D Total
88 76 130 71 105 84 110 64 728
At this point the checks and the check effects are estimated followed by progeny
effect as shown in Tables 4.64 and 4.65.
1. Compute check effects: Let C = check total, c = check effect and cC = the product
of the check total and check effect for each checks.
73
Biometry
1
= [2889 -(4-1)(293.5) – 0.5] c = number of checks
4 + 16
p = number of progenies
= 100.4 b = number of blocks
74
Biometry
8. Crude block SS =
∑ Bi
ni
[(690) 2 + (728) 2 + (811) 2 + (660) 2
=
8
=262425.63
9. True block SS (TBSS)
2
a. Ignoring entries =
∑ Bi
– C.F.
ni
= 262425.63 – 260822.53
= 1603.1
b. Eliminating = TBSS – [SSeib– SSeeb] where eib= entries ignoring blocks
eeb = entries eliminating block
= 1603.10 – [15776.97 – 14192.7]
= 18.83
Table 4.66. Adjusted progeny mean
75
Biometry
11. SSeib =
∑ Ci + Crude SS progenies - (C.F.)
b
[(120) 2 + (100) 2 + ... + (103) 2 ]
= – 260822.53
4
= 15776.97
2 2
12. SS checks =
∑ Ci – CF’ (of the checks), where C.F.’ = ∑ Ci = (1174) 2 = 86142.25
b cb 4 x4
[(335) 2 + (306) 2 + (278) 2 + (255) 2 ]
= – 86142.25
4
= 900.25
2
2
13. SS progenies = ∑Pj – CF’’ (of the progenies), C.F.’’ =
∑ Pi
=
(1715) 2
= 183826.56
P 16
= [(120)2 + (100)2 + … + (103)2] – 183826.56
= 5730.44
14. SS check x progenies = SSeib – SS checks – SS progenies
= 15776.97 – 900.25 – 5730.44
= 9146.28
15. SS error = Total SS – TBSS ignoring - SSeeb
= 15798.47 – 1603.10 – 14192.70
= 2.67
76
Biometry
Entries 19
a. eliminating blocks 14192.70 746.98 2489.93 2.96
b. ignoring blocks 15776.97 830.37 2767.90 2.96
77
Biometry
1
= 2 x0.3(1 + )
4
= 0.87 where c = no. of checks
CD = SE(d) x t0.05 at error D.F.
= 0.87 x 2.262
= 1.97
d. To compare two progenies vs. checks
1 1 1
SE (d) = MS E (1 + + + )
b c bc
1 1 1
= 0.3(1 + + + )
4 4 4 x4
= 0.68
78
Biometry
D 63.75
Interpretation
There was a significant difference among progenies, checks and progenies vs.
checks. Among the progenies, P10 had the highest yield (120 g/plot) followed by P12
and P11 while the least was obtained from P4 (but P4and P13 are statistically not
different). Among the checks, the highest yield (83.75 g/plot) was obtained from A
followed by B while the least was obtained from D. When progenies and checks were
compared, the highest yield was obtained from P10 followed by P12. However,
progeny 4 and check A were not statistically different. Therefore, all progenies except
P4 can be promoted for further test based on the yield per plot. P4 can only be
considered if it has special feature like disease or insect resistance.
Incomplete block designs are types of designs where the experimental units are
grouped into blocks that cannot contain a full replication of all treatments. One of the
basic principles of modern experimental design is to group the experimental units into
blocks of homogeneous units to increase precision and to make comparisons under
uniform conditions. In general, precision increases as the block size decreases and
hence, smaller blocks are preferred to larger ones. The disadvantages of incomplete
block designs are: restricted number of treatments and replications or both, unequal
degree of precision in comparing treatment means and complex data analysis.
The incomplete block designs are categorized into two: balanced designs and partial
balanced designs. In balanced designs each treatment occurs together in the same
block with every other treatment equal number of times, usually once. In a balanced
incomplete block design all pairs of treatments are compared with the same precision
even if the differences between blocks may be large.
The balanced lattice designs are restricted in such a way that the number of
treatments must be a perfect square, the block size (k) is equal to the square root of
the number of treatments and the number of replications is one more than the block
size (k+1). In such cases, the minimum number of replication is fixed to balance and
that makes the balanced design less practical. As a result of this, a class designs
called partially balanced incomplete block designs were developed.
79
Biometry
Among the most commonly used incomplete block designs, lattice designs are called
fore. In lattice designs the number of treatment must be a perfect square. The
number of blocks and the number of plots in each block is the square root of the
number of treatments. In lattice design the incomplete blocks are combined into
groups that form separate, complete replications. When a lattice design consists of
two or more complete replications it has been shown that the design can be analyzed
as if it were an ordinary randomized complete block design (Cochran and Cox, 1957).
Hence the blocking restriction can be ignored without destroying the validity of the
analysis even thought the precision may be reduced.
In all the lattice designs the number of treatment must be a perfect square, and the
number plots per incomplete block is the square root of the number of treatments.
The designs differ from each other primarily by the number of replications. Field plan
for selected lattice designs are given in Appendix 8.
The blocks in lattice designs should be composed of plots that are as homogeneous
as possible. Blocks in the same replication should also be as uniform as possible to
maximize the variation among replications. This can maximize precision if the
experiment is analyzed as a randomized complete block design. Lattice designs can
be categorized into two classes: balanced lattice and partially balanced lattice.
In balanced lattice, the number of treatments must be a perfect square and the size
of block is the square root of the number of treatments. The special feature of the
balanced lattice, as distinguished from other lattices, is that every pair of treatments
occurs once in the same incomplete block. Consequently, all pairs of treatments are
compared with the same degree of precision (Cochran and Cox, 1957).
In balanced lattice each treatment occurs together with every other treatment in the
same block equal number of times. However, if there are b number of blocks, there
must be b + 1 number of replication to achieve balance. This restriction in the number
of replication and treatments makes the design less practical in terms of resources.
Balanced lattice design cannot be constructed for 36, 100 and 144 number of
treatments.
4.6.2.1. Layout
For example, in an experimental field where there is a high variation that runs in more
than one direction, RCBD cannot be used because the high variability in the field can
80
Biometry
introduce within block variation. It is also not possible to use Latin square design as it
requires larger experimental area that definitely minimizes our precision. The next
option is to seek for a suitable experimental design. Lattice designs are suitable in
such cases. Suppose there are nine treatments that need bigger plot size in a site
where the experimental area has got fertility variation which can not be controlled
through blocking in one direction (as in RCBD) or in two directions (Latin SD as it
makes the experimental area very large, 9 x 9 LSD). Hence, the experiment is
planned as a 3 x 3 balanced lattice design so that the fertility difference can not
contribute to intra-block error. The treatments were placed into three blocks each
containing three treatments as shown in Figure 4.7.
4.6.2.2. Analysis
1. Calculate the block totals, the replication totals, the grand total and the treatment
total as shown in Table 4.69.
81
Biometry
Grand total GT
1. For each treatment, Bt = sum of all blocks for a treatment containing that a
particular treatment will be calculated. For example, for treatment 1,
Bt = B1 + B4 + B7 + B10
2. The quantities, W = kT - (k+1)Bt +GT, where ∑W = 0.
3. Total SS, SSt, SSR are computed as usual and the SSB within replications,
The symbol for adjustment factor (A) can also have different symbols, for example
Cochran and Cox (1957) used µ but the most important thing is to understand the
meaning. In this manual as mentioned in the introductory part, we use notations that
are easy to pick, like the first letter of the term (example, Adjustment factor = A,
Grand total = GT, grand mean = GM, Correction factor = CF and so on) and
notations used by most authors like µ. If Eb < Ee, A is taken as zero and no
adjustment is applied to the treatment totals.
5. For t-test, calculate the effective error mean square, E’e = Ee(1+kA). The variance
of the difference between two adjusted treatment totals is 2r E’e, while for the
difference between two adjusted treatment means is
2 E 'e
.
r
The treatment mean square in the above case (Table 4.70) cannot be tested against
the intra-block error mean square because the treatment mean square contains
82
Biometry
some block effects. Rather the adjusted treatment mean square is tested against the
effective error mean square, E’e
83
Biometry
Table 4.71. Grain yield data (t/ha) of 16 hybrids evaluated in 4 x 4 balanced lattice design
Rep
Trt I II III IV V
b1 b2 b3 b4 b5 b6 b7 b8 b9 b10 b11 b12 b13 b14 b15 b16 b17 b18 b19 b20 Tj Bj Wj
T1 14.9 14.0 15.0 22 14.8 80.7 329.9 55.8
T2 15.2 15.6 13.2 15.9 20.5 80.4 336.2 23.1
T3 16.7 19.2 15.3 20.5 19.6 91.3 346.2 16.7
T4 15.6 16.2 19.5 17.6 22.6 91.5 361.2 -57.5
T5 12.7 16.0 18.0 16.8 17.8 81.3 343.7 -10.8
T6 15.5 14.2 16.5 17.8 21.3 85.3 333.0 58.7
T7 16.2 15.6 18.5 17.8 19.6 87.7 359.2 -62.7
T8 17.2 20.9 15.9 18.9 14.8 87.8 334.4 61.3
T9 14.3 18.0 13.9 15.6 18.0 80.2 332.3 41.8
T10 10.0 9.8 12.5 15.6 11.5 59.4 316.6 37.1
T11 19.2 19.5 17.9 18.6 20.6 95.8 356.2 -15.3
T12 17.0 21.2 16.6 20.8 19.5 95.1 362.4 -49.1
T13 15.5 15.2 18.5 20.8 17.5 87.5 358.0 -57.5
T14 19.0 19.3 18.6 20.5 22.5 99.9 360.9 -22.4
T15 19.2 18.2 17.9 18.5 17.0 90.8 339.5 48.2
T16 20 18.5 15.7 16.8 16.9 87.9 360.3 -67.4
Bi 62.4 61.6 60.9 73.7 63.2 58.9 72.5 76.8 65.1 65.7 63.7 69.0 81.1 74.2 69.7 69.5 58.1 75 77.9 83.5 1382.5 5530 0
Rk 258.6 271.4 263.5 294.5 294.5
84
The stepwise analysis is as follows:
2. Compute Bj
a. B1 = Bi1 + Bi5 + Bi9 + Bi13 + Bi17
= 62.4 + 63.2 + 65.1 + 81.1 + 58.1
=329.9
b. B2 = Bi1 + Bi6+ Bi10 + Bi14 + Bi18
= 62.4 + 58.9 + 65.7 + 74.2 + 75
=336.2
c. B3 = Bi1 + Bi7 + Bi11 + Bi15 + Bi19
= 62.4 + 72.5 + 63.7 + 69.7 + 77.9
=346.2
d. B4 = Bi1 + Bi8+ Bi12 + Bi16 + Bi20
= 62.4 + 76.8 + 69.0 + 69.5 + 83.5
=361.2
e. B5 = Bi2 + Bi5 + Bi10 + Bi15 + Bi20
= 61.6 + 63.2 + 65.7 + 69.7 + 83.5
=343.7
f. B6 = Bi2 + Bi6 + Bi9 + Bi16 + Bi19
= 61.6 + 58.9 + 65.1 + 69.5 + 77.9
=333.0
g. B7 = Bi2 + Bi7 + Bi12 + Bi13 + Bi18
= 61.6 + 72.5 + 69.0 + 81.1 + 75.0
=359.2
h. B8 = Bi2 + Bi8 + Bi11 + Bi14 + Bi17
= 61.6 + 76.8 + 63.7 + 74.2 + 58.1
=329.9
i. B9 = Bi3 + Bi5 + Bi11 + Bi14 + Bi17
= 60.9 + 63.2 + 63.7 + 74.2 + 58.1
=332.3
j. B10 = Bi3 + Bi6 + Bi12 + Bi15 + Bi17
= 60.9 + 58.9 + 69 + 69.7 + 58.1
=329.9
k. B11 = Bi3 + Bi7 + Bi9 + Bi14 + Bi20
= 60.9 + 72.5 + 65.1 + 74.2 + 83.5
=356.2
l. B12 = Bi3 + Bi8 + Bi10 + Bi13 + Bi19
= 60.9 + 76.8 + 65.7 + 81.1 + 77.9
=362.4
m. B13 = Bi4 + Bi5 + Bi12 + Bi14 + Bi19
= 73.7 + 63.2 + 69 + 74.2 + 77.9
=358
n. B14 = Bi4 + Bi6 + Bi11+ Bi13 + Bi20
= 73.74 + 58.9 + 63.7 + 81.1 + 83.5
=360.9
Biometry
86
Biometry
4. Compute Sum of squares for the different components. The best way is to
start with the adjusted block sum of squares because the mean square of the
block is an important component for making decision whether we continue the
analysis as lattice or as RCBD after comparing it with MS of error.
2
a. SS block (adjusted) = 3
∑ Wj
q (q + 1)
109.14
=
42 − 1
= 7.28 = Eb
(GT ) 2 (1382.5) 2
c. Correction factor(C.F.) = = = 23891.33
rq 2 5 x 42
d. Total SS = ∑Xijk2 – CF = [(14.9)2 + (15.2)2 + … + (22.5)2] - 23891.33
= 566.4
e. SS treatment (unadj.)
2
SSt (unadj.) =
∑ Ti
− C.F .
r
(80.7) + (80.4) 2 + ... + (87.9) 2 ]
=
5
= 257.13
f. SS replication (SSR)
SSR =
∑ R j 2 − C .F .
q2
87
Biometry
Eb − Ee 7.28 − 2.83
j. µ = 2
= = 0.04, since µ is positive we will proceed to adjust
q Eb 4 2 x 7.28
treatment means as in Table 4.72.
5. SSt (adjusted) =
∑T i
– CF
q +1
[(82.93) 2 + (81.32) 2 + ... + (85.20) 2 ]
= - 23891.33
4 +1
88
Biometry
= 223.77
6. MSt = SSt(adj)/q2 –1= 223.77/42-1 = 14.92
7. Effective error MS = Ee (1 + qµ) = 2.83(1+4x0.04) = 3.2828
8. F-calculated = MSt/Effective error MS = 14.92/3.2828 = 4.54
9. Finally ANOVA table can be constructed as in Table 4.73
(109.14) + (127.42)
=
15 + 45
= 3.94
89
Biometry
_________________________________
T14 19.80
T11 19.04
T12 18.63
T15 18.55
T3 18.39
T8 18.03
T4 17.84
T6 17.53
T16 17.04
T13 17.04
T7 17.04
T1 16.59
T9 16.37
T2 16.26
T5 16.17
T10 12.18
_________________________________
CD/LSD 2.54
c.v. 11.56%
_________________________________
Interpretation
Interpretation
Treatment effect was found to be significant (Table 4.73). T14 had the highest grain
yield (19.80 t/ha) followed by T11 and T12 (but statistically these three treatments did
not show differences among themselves) as shown in Table 4.74. There was also a
significant block effect implying that blocking helped in reducing experimental error.
The relative efficiency of 1.20 indicates that the use of lattice design instead of RCBD
improved precision by 20%.
The names of the sub categories of partially balanced design follow the number of
replication. For example, the balanced lattice with two replications is called simple
lattice, with three replication triple lattice, with four replications quadruple lattice and
so on. In such arrangement, some treatment pairs never appear together in the same
90
Biometry
incomplete block (λ = 1 or λ = 0). That is why treatment means are compared with
different level of precision, as there is more than one level of precision.
Let us assume that the experimental area has variability with unknown direction that
RCBD could not control the variability. We cannot use Latin square as it requires
larger experimental area that definitely minimize our precision. The resource we
have is not enough to use a balanced lattice. The next option is to use the simple
lattice design as it fulfills the requirements.
For 9 and 16 treatments, simple lattices are unlikely to be more accurate than RCBD
unless the variation among incomplete blocks is great compared with that within
incomplete blocks (Cochran and Cox, 1957). Besides, the degrees of freedom are
lower in simple lattices than RCBD.
Cochran and Cox, (1957) recommend to analyze the data as RCBD if missing
observations are numerous. The full analysis can also be done by estimating the
missing values by the formula, which minimizes the intra-block error sum of squares
as shown below. Hence, simple lattice design is suggested for use where the number
of treatments ≥ 16.
4.6.3.2.1. Layout
Suppose there are 16 treatments that need bigger plot size in a site where the
experimental area has got variation in soil acidity with unknown gradient which
cannot be controlled through blocking in one direction (as in RCBD) or in two
directions (Latin SD as it makes the experimental area very large, 16 x 16 LSD).
Therefore, the experiment is planned as a 4 x 4 simple lattice design so that the soil
acidity difference cannot contribute to intra-block error. The treatments are placed
into four blocks each containing four treatments as shown in Figure 4.8.
91
Biometry
4.6.3.2.2. Analysis
The stepwise analysis of variance for t number of treatments, k number of blocks and
r number of replication is as follows.
1. The totals for blocks, replications and treatments are calculated as shown in Table
4.75.
Table 4.76. Summary of treatment total and adjusted means of simple lattice design
Total Mean
Treatment Unadjusted (Ti) Adjusted (TAi) Adjusted (AM)
1 T1 TA1 MA1
2 T2 TA2 MA2
. . . .
. . . .
. . . .
q Tq2 TAq2 MAq2
92
Biometry
Ti = sum of the ith treatment, TA = treatment total adjusted (Ti + ACil over all blocks where the
ith treatment appear) and AM = adjusted total/number of replication.
The sum of squares for total, replication, treatment and error are computed as in any
other designs. The sum of squares due to block is a new statistic to be computed in
lattice designs.
(GT ) 2
1. Correction factor C.F. =
rq 2
2. Total SS = ∑∑X2ij(l) – C.F.
2
3. SSR =
∑ Rj
–C.F.
q2
2 2
4. SSB =
∑∑ C ij
-
∑C i
qr (r − 1) q 2 r (r − 1)
2
5. SSt =
∑T i
– C.F.
r
6. SSE = Total SS – SSR – SSB – SSt
Sources d.f. SS MS
of variation
Replications (r-1) SSR
Treatments (t-1) SSt
Blocks (adj.) r(q-1) SSB Eb
Intra block error (q-1)(rq-q-1) SSE Ee
8. The mean square of block and error are computed as usual dividing the sum of
squares of block and error by their respective degrees of freedom:
93
Biometry
SS B
Eb = and
r (q − 1)
SS E
Ee =
(q − 1)(rq − q − 1)
Then these two mean squares are compared either to go for adjustment factor or not.
If Eb ≤ Ee, the adjustment for block has no effect. That will lead us to ignore the
blocking restriction and analyze the data as if the design had been a randomized
block design with replications as blocks. If on the other hand, Eb> Ee, an adjustment
factor, µ, is computed for the design.
E − Ee
µ= b
qEb
This adjustment factor is used to compute ACil in Table 4.75 and adjusted treatment
means in the table of total, Table 4.76.
Finally, the effective error mean square that can be used in calculating t-test and
interval estimates is calculated as:
1 + rqµ
E’e = .
(q + 1) Ee
Adjusted treatment mean square is computed to test whether there is a significant
difference among adjusted treatment means or not. In order to do that, it is necessary
to compute first the unadjusted blocks within replications sum of squares (SSBu).
2
SSBu =
∑∑ Bil
– C.F. – SSR.
q
Then the adjusted treatment sum of squares, SSt(adj.), is computed as
2
SSt(adj.) = SSt(unadj.) – µq( )[SSBu/(r-1)(1+µq) – (SSB(adj.))].
1 + qµ
SSt
MSt =
t −1
MSt
F=
Ee
2 Ee
SE(d)1 = [1 + (r − 1) µ ] , is used for comparing treatments within a block
r
2 Ee
SE(d)2 = [1 + rµ ] , is used for comparing treatments in different blocks
r
94
Biometry
Ee 1 + 2kµ
B. Effective MSE = E’e = [ ].
r (k + 1)
A
Relative efficiency (R.E.) of simple lattice over RCBD =
B
% efficiency = (R.E. – 1)100
Block Rep-I BK
1 650 (1) 670(2) 720(3) 680(4) 2720
2 685(5) 655(6) 670(7) 908(8) 2918
3 685(9) 680(10 680(11) 685(12) 2730
4 725(13) 690(14) 735(15) 685(16) 2835
CT1 2745 2695 2805 2958 11203
R1
Block Rep-II BK
5 735(1) 690(5) 840(9) 805(13) 3070
6 650(2) 685(6) 820(10) 735(14) 2890
7 670(3) 675(7) 685(11) 745(15) 2775
8 910(4) 790(8) 825(12) 855(16) 3380
CT2 2965 2840 3170 3140 12115
R2
GT = R1 + R2 = 11203 + 12115 = 23318
BK= block total, ( ) = treatment, CT1 = column total of rep-I, = column total of rep-II, xij = ith treatment
in jth block GT1 grand total of rep-I, GT2 = grand total of rep-II, R1 = total of rep-l, R2 = total of rep-II
and GT = grand total.
95
Biometry
T7 = 1345
T8 = 1698
T9 = 1525
T10 = 1500
T11 = 1365
T12 = 1510
T13 = 1530
T14 = 1425
T15 = 1480
T16 = 1540
2. Compute correction term for each block
B’1 = (B1 + C5) – rB1
= (2720 + 2965) – 2 x 2720
= 245 or = (T1+T2+T3+T4)-rB1
B’2 = (B2 + C6) – rB2
= (2918 + 2840) – 2 x 2918
= -78
B’3 = (B3 + C7) – rB3
= (2730 + 3170) – 2 x 2730
= 440
B’4 = (B4 + C8) – rB4
= (2835 + 3140) – 2 x 2835
= 305
B’5 = (B5 + C1) – rB5
= (3070 + 2745) – 2 x 3070
= -325
B’6 = (B6 + C2) – rB6
= (2890 + 2695) – 2 x 2890
= -195
B’7 = (B7 + C3) – rB7
= (2775 + 2805) – 2 x 2775
= 30
B’8 = (B8 + C4) – rB8
= (3380 + 2958) – 2 x 3380
= -422
96
Biometry
= 912 + (-912)
=0
(GT ) 2 (23318) 2
a. Correction factor C.F. = = = 16991535.13
rq 2 2 x 42
b. Total SS = ∑∑X2ijk – C.F.
=(650)2 + (685)2 + … + (855)2 – 16991535.13
=173628.87
SSR =
∑ R j – C.F.
q2
(11203) 2 + (12115) 2
= – 16991535.13
42
=25991.995
d. Sum of squares due to block(unadjusted (SSB(unadj)
2
SSB (unadj) =
∑ Bk
– C.F.
q
(2720) 2 + (2918) 2 + ... + (3380) 2
= – 16991535.13
4
=84783.37
e. Sum of squares due to treatment (SSt)
2
SSt =
∑ Ti
– C.F.
r
(1385) 2 + (1320) 2 + ... + (1540) 2
= – 16991535.13
2
=84341.87
4. The SSB (adj), d.f. for block , MSB and Eb
2 2
SSB(adj) =
∑ B' k
–
∑ R' j
kr (r − 1) k 2 r (r − 1)
= [(B’12 + B’22 + . . . + B’82)/4x2(2-1)] - [(R’12 + R’22)/42x2(2-1)]
(245) 2 + (−75) 2 + ... + (−422) 2 (912) 2 + (−912) 2
= -
4 x 2(2 − 1) 42 x 2(2 − 1)
= 32437
5. Degree of freedom for block = 2(k-1) = 2(4-1) = 6
97
Biometry
SS B adj 32437
MSB(adj.) = = = 5406.17
block d . f . 6
6. The error sum of squares, degrees of freedom and mean square
SSE = Total SS – SSR – SSt(unadj.) – SSB(adj.)
= 173628.87 – 25991.995 – 84341.87 – 32437
= 30858.01
7. d.f. for error = (q-1)(2q-q-1) = (4-1)(2x4 – 4 – 1) = 9
SS E
MSE =
error d . f .
30858.01
=
9
= 3428.67
8. Compute µ
Eb − Ee
µ=
qEb
5406.17 − 3428.67
=
4(5406.17)
= 0.09
ns = non-significant
c.v. = 8.04
98
Biometry
In the practical situations, the analysis is terminated at this point because the
treatment effect was not significant. As the manual is intended to show the statistical
methods related to this particular design, the standard error of differences and
adjusted treatment means will be computed to make the statistical procedure
complete. The remaining component is to compare the treatments following the
adjusted treatment means.
99
Biometry
2 Ee
SE(d)1 = [1 + (r − 1) µ ] , (for comparing treatments within a block)
r
2 x3428.67
= [1 + (2 − 1)0.09]
2
= 61.13
2 Ee
SE(d)2 = [1 + rµ ] , (for comparing treatments in different blocks)
r
2 x3428.67
= [1 + 2 x0.09]
2
= 63.61
100
Biometry
Interpretation
The effect due to block was significant indicating that blocking in lattice manner was
effective in controlling the variability of the soil acidity (helped in reducing
experimental error). The relative efficiency was 115% implying that the use of lattice
design increased precision by 115% compared to RCBD. The treatment effect was
not significant.
The main-plot factor effects are estimated from the large units, while the sub-plot
factor effects and the interaction of the sub-plot factor with the main-plot are
estimated from the sub-plots. There are two experimental errors, one coming from
the main-plot and the other one from the sub-plot, as they differ in size of the plots.
Generally the main-plot error is larger than the sub-plot error and as a result the
main-plot factor is estimated with less precision than the factor in the sub-plot and the
interaction. Hence, it is important to consider the difference in precision at the time of
assigning factors into either of the plots. Gomez and Gomez (1984) suggested the
following guidelines in assigning factors into either main- or sub-plot: degree of
precision, relative size of the main effect and management practices.
101
Biometry
The advantages of this design is that it permits the efficient use of some factors that
require different plot size, permits introduction of new treatments into an on-going
experiment and also provides increased precision in the estimation of some of the
factorial effects.
The disadvantages of the design are the less precision associated with the main-plot
(which requires large differences for significance) and that the statistical analysis is
more complex compared to designs for single factor experiments.
5.2.2. Lay-out
Two independent randomizations are necessary in this design. The first
randomization is done for the main-plot and the second for the sub-plot. In both
cases, the use of random number is important. Let the number of the main plot
treatment is denoted by a, the number of sub-plot treatments by b and the number of
replication by r. Then, there will be abr number of plots. Other components remain as
in RCBD.
5.2.3. Analysis
The analysis of variance for split-plot design involves two steps. The first step is main
plot analysis and the second step is the sub plot analysis. For example, if there are
two factors: factor A at a levels as a main plot and factor B at b levels replicated r
times. The computation will be done as follows:
(GT ) 2
1. Correction factor (C.F.) = , where a is the number of levels of A
abr
b is the number of levels of B
r is the number of replication
2. Total sum of squares (Total SS)
Total SS =∑∑∑Xijk 2 –C.F.
SSR =
∑ Rk
– C.F.
ab
4. Sum of squares due Main-plot (SSA). Here we need to have a two-way table
between main-plot and replication (Table 5.1).
102
Biometry
Then, SSA =
∑A i
– C.F.
br
SSB =
∑ Bj
– C.F.
ar
103
Biometry
The statistical objectives of this exercise are to get familiarized with the analysis of
variance for split-plot, to identify the important factor, optimum level of each factor
and combination of factors, and to investigate whether there is interaction effect
between the two factors or not.
The biological objective is to investigate the optimum sowing time and plant density, if
factors are interdependent to identify best combination of the factors that can give
highest yield.
104
Biometry
The hypothesis of the experiment is that the different levels of each factors act
independently and the alternative hypothesis at least two levels of the factors
interact. The data collected from the experiment was provided in Table 5.5 for
analysis.
Table 5.5. Grain yield (t/ha) of a wheat variety at different density and sowing time
Replication
Sowing date Total
I II III IV V VI Sowing
Density S1 Date(Si)
D1 5.59 5.50 5.25 5.31 4.63 4.69
D2 5.69 6.66 6.22 5.84 5.97 6.03
D3 6.90 5.28 6.66 6.19 6.40 6.38
Sum 18.18 17.44 18.13 17.34 17.0 17.1 105.19
S2
D1 5.84 5.63 6.14 5.94 6.13 6.08
D2 6.90 6.86 7.03 6.94 6.22 6.88
D3 7.09 6.72 6.25 6.36 6.34 6.81
19.83 19.21 19.42 19.24 18.69 17.77 116.16
S3
D1 5.23 5.41 5.25 5.71 5.06 5.75
D2 6.60 6.55 7.00 6.28 6.75 7.03
D3 6.03 6.52 6.12 5.72 6.13 6.88
Sum 17.86 18.48 18.37 17.71 17.94 19.66 110.02
Let the main plot is denoted by A, the sub plot by B and the number of replication by
r. Then, the stepwise analysis of variance for Split-plot with 6 replications will be done
as follows.
1. The linear additive model (L.A.M.): Xijk = µ + Ai, Bj, (AB)ij + Rk + Eijk where, Xijk =
ijkth observation, µ = the overall mean, Ai = the ith level of factor A, Bj = the jth
level of factor B, (AB)ij = ijth level of combination(interaction of factor A and B, Rk =
kth replication effect and Eijk = the error associated with ijkth observations.
Assumptions
Both factors (Density and sowing date) ~ NI (0, σ2t)
Rj ~ NI (0, σ2R)
Eijk ~ NIR (0, σ2)
(GT ) 2
2. Correction factor (C.F.) = , where a is the no. of levels of A
abr
(331.37) 2
= b is the no. of levels of B
3 x3 x 6
r is the no. of replication
105
Biometry
=2033.45
3. Total sum of squares (Total SS)
Total SS = ∑∑∑Xijk 2 – C.F.
= [(5.59)2 + (5.50)2 + … + (6.88)2] – 2033.45
=21.00
4. Sum of squares due to replication (SSR)
2
SSR =
∑ Rk
– C.F.
ab
(55.87) 2 + (55.13) 2 + ... + (56.53) 2
= – 2033.45
3 x3
= 0.67
5. Sum of squares due sowing date (SSS)
Here we need to have a two-way table between main-plot and replication(Table 5.6).
Table 5.6. Grain yield (t/ha) of a wheat variety at different density and sowing time
S\rep I II III IV V VI Si
S1 18.18 17.44 18.13 17.34 17.0 17.1 105.19
S2 19.83 19.21 19.42 19.24 18.69 17.77 116.16
S3 17.86 18.48 18.37 17.71 17.94 19.66 110.02
Rk 55.87 55.13 55.92 54.29 53.63 56.53 331.37
Then, SSS =
∑S i
– C.F.
br
(105.19) 2 + (116.16)2 + (110.02) 2
= – 2033.45
3 x6
= 3.35
5. Sum of squares due to Error (a)
Total SS ( SxR)
SSE(a) = – C.F. – SSS - SSR
b
(18.18) 2 + (17.44) 2 + ... + (19.66) 2
= – 2033.45 - 0.67 - 3.35
3
= 0.91
6. Sum of squares due Density (SSD)
It is necessary to make a two-way table between sowing time and density as Table
5.7.
Table 5.7. Grain yield (t/ha) of a wheat variety at different density and sowing time
S\rep D1 D2 D3 Si
S1 30.97 36.41 37.81 105.19
S2 35.76 40.83 39.57 116.16
S3 32.41 40.21 37.40 110.02
Dj 99.14 117.45 114.78 331.37
106
Biometry
SSD =
∑D j
– C.F.
ar
(99.14) 2 + (117.45) 2 + ... + (114.76) 2
= – 2033.45
3x6
= 10.87
7. Sum of squares due sowing time x Densities (SSSxD)
Total SS ( SxD)
SS SxD = – C.F.
r
(30.97) 2 + (36.41) 2 + ... + (37.40) 2
= – 2033.45
6
= 1.01
8. Sum of squares due to error (b)
SSE (b) = Total SS – SSR – SSS - SSE(a) – SSD - SS SxD
= 21.00 – 0.67 – 3.35 – 0.91 – 10.87 – 1.01
= 4.19
9. Degrees of freedom for replication (d.f. for R)
d.f. for R = number of replication (r)-1
= 6-1
=5
10. Degrees of freedom for sowing time (d.f. for S)
d.f. for S = (a-1)
= (3-1)
=2
11. Degrees of freedom for error(a)
d.f. for Ea = (r-1)(a-1)
= (6-1)(3-1)
= 10
12. Degrees of freedom for plant density (d.f. for D)
d.f. for D = b-1
= 3-1
=2
13. Degrees of freedom for sowing time x Density (d.f. for IxD)
d.f. for IxD = (a-1)(b-1)
= (3-1)(3-1)
=4
14. Degrees of freedom for error (b) (d.f. for Eb)
d.f. for Eb = a(b-1)(r-1)
= 3(3-1)(6-1)
= 30
15. Mean squares
SS of each component divided by the corresponding degree of freedom
as in the ANOVA Table 5.8.
16. Calculate F-values
107
Biometry
Then the calculated F-vales are compared with F-tabulated at 0.05 and 0.01 level of
probability: (10, 30) = 2.16 and 2.98. Error (a) has F-calculated less than F-tabulated,
i.e. not significant, and hence can not be used for testing the effect of replication and
sowing time. Because non-significant F-calculated indicates that the mean square
error (a) is zero and dividing any number by zero is not mathematically logical.
The interpretation when mean square error (a) is non significant is that our
consideration during planning of assigning the sowing time as main-plot was not
effective in controlling the error associated with the main plot. You may find some
literature using error (a) without testing but at this level it is important to be
considerate enough with this kind of apparent confusion.
Table 5.8. ANOVA table for sowing time x plant density study in split-plot design
The analysis of variance showed that the effect of sowing time and plant density were
significant but the interaction effect was not-significant implying that the two factors
acted independently in this particular experiment. Hence the best level of each factor
can be identified by computing the critical value for each factor.
108
Biometry
0.1397
=±
3 x6
= ±0.09
2 MS E ( b )
SE(d) for all components =±
ar
2 x0.1397
=±
3x6
= ±0.13
CD(LSD) = SE(d) x t0.05 at error d.f.
= 0.1397 x 2.042
= 0.27
Table 5.9. Mean yield (t/ha) of different sowing date and density on wheat
109
Biometry
2. Density
D-1 5.51 0.27 6.10
D-2 6.53
D-3 6.38
Interpretation
The analysis of variance for split-plot design (Table 5.8) showed that the effects due
sowing dates and spacing were significant while the interaction was not indicating
that the two factors acted independently. The MS due to error (a) was not significant
indicating that assigning sowing date as the main plot was not correct.
The highest yield was obtained at sowing date (S-2) while the least was due to check
(S-1). Likewise, the highest yield was obtained at density (D-2). Hence, S-2 and D-2
can be recommended for the area where the experiment was conducted (Table 5.9).
Since there are three plot sizes, there will be three experimental errors, one for each
plot size. The interaction is measured with greater precision than the main effects.
The principal advantage of the split-block design is that it permits the efficient
application of factors that would be difficult to apply on small plots. It is also useful to
introduce new factor into an ongoing experiment. The disadvantages of this design
are similar to split-plot in that it has unequal precision (in estimating the interaction
and the main effects) and complicated statistical analysis.
5.2.1.Lay-out
In a split-block design, the levels of one factor are assigned to strips of plots running
through the block in one direction with a separate randomization in each block. The
levels of the second factor are then applied to strips of plots that are oriented
perpendicular to the strips for the first factor. The field plan for any number of
replication of two factors can be done as in RCBD, separate randomization for each
factor. Details can be obtained from (Gomez and Gomez, 1984).
5.2.2.Analysis
The analysis of data from split-block design follows the same general pattern as any
other design. For example, if there are two factors, i.e., factor A at a levels assigned
at random to a strip of b plots each and factor B at b levels to b strips of a plots each
110
Biometry
crossing the strips used for the A factor. If the treatments are replicated r times and
Xijk represents an observation of the ith level of factor A, jth level of factor B in kth
replication, then the data can be summarized as in Tables 5.18. As there are three
plots, there will be three summary tables (Tables 5.18 to 5.20).
(GT ) 2
1. Correction factor (C.F.) = , where a is the no. of levels of A
abr
b is the no. of levels of B
r is the no. of Replication
2. Total sum of squares (Total SS)
Total SS = ∑∑∑Xijk 2 –C.F.
111
Biometry
SSR =
∑ Rk – C.F.
ab
4. Sum of squares due Main-plot (SSA). Here we need to have a two-way table
between main-plot and replication (Table 5.18).
SSA =
∑A i
– C.F.
br
5. Sum of squares due to Error (a)
Total SS ( AxR)
SSE(a) = – C.F. – SSA - SSR
b
SSB =
∑ B j 2 – C.F.
ar
112
Biometry
Table 5.22. Yield of tomato as affected by green manure and date of transplanting
Replication
I II III IV Gi
g1 d1 23.7 25.5 28.0 27.2
d2 40.8 42.3 48.8 49.2
d3 44.5 41.6 45.8 43.1
d4 25.7 27.2 28.2 26.6
Sub total 134.7 136.6 150.8 146.1 568.2
g2 d1 30.1 29.2 33.5 31.5
d2 45.6 48.3 50.2 46.6
d3 48.0 52.1 55.5 50.2
d4 29.2 31.5 30.2 53.5
Sub total 152.9 161.1 169.4 181.8 665.2
g1 d1 35.2 38.4 39.5 40.2
d2 50.5 52.6 58.5 54.5
d3 54.1 53.6 59.5 56.2
d4 32.0 31.8 34.0 32.8
113
Biometry
1. The linear additive model (L.A.M.): Xijk = µ + Ai, Bj, (AB)ij + Rk, + Eijk
where, Xijk = ijkth observation, µ = the overall mean, Ai = the ith level of factor A, Bj
= the jth level of factor B, (AB)ij = ijth level of combination(interaction of factor A
and B, Rk = kth replication effect and Eijk = the error associated with ijkth
observations.
Assumptions
Factors A and B ~ NI (0, σ2t)
Rj ~ NI (0, σ2R)
Eijk ~ NIR (0, σ2)
(GT ) 2
3. Correction factor (C.F.) = , where g= number of levels of G and
gdr
(2797.7) 2
= d = number of levels of D
4 x4 x4
=122298.83
4. Total sum of squares (Total SS)
Total SS = ∑∑Xijk 2 –C.F.
=(23.7)2 + (40.8)2 + … + (39.2)2 – 122298.83
=9558.08
114
Biometry
Replication
Green manure I II III IV Gi
g1 134.7 136.6 150.8 146.1 568.2
g2 152.9 161.1 169.4 181.8 665.2
g3 171.8 176.4 191.5 183.7 723.4
g4 201.8 211.6 217.7 209.8 840.9
Rk 661.2 685.7 729.4 721.4 2797.7
Total SSGR =
∑∑ X i.k
– C.F.
d
(134.7) 2 + (136.6) 2 + ... + (209.8) 2
= – 122298.83
4
= 2682.67
SSR =
∑ Rk
– C.F.
gd
SSG =
∑ Gi 2 – C.F.
dr
(568.2) 2 + (665.2) 2 + ... + (840.9) 2
= – 122298.83
4 x4
= 2436.34
8. Sum of squares due to Error green manure (g)
Total SSGxR
SSE(a) = – SSG - SSR
d
= 2682.67 – 2436.34 – 189.44
= 56.89
115
Biometry
Replication
Green manure I II III IV Dj
d1 127.3 133.3 143.5 140.4 544.5
d2 197.1 206.3 221.5 211.8 836.7
d3 211.9 215.4 230.0 217.1 874.4
d4 124.9 130.7 134.4 152.1 542.1
Rk 661.2 685.7 729.4 721.4 2797.7
Total SS DxR
Total SSDR = – C.F.
g
(127.3) 2 + (133.3) 2 + ... + (152.1) 2
= – 122298.83
4
= 6406.08
10. Sum of squares due to replication (SSR) the same as above = 189.44
SSD =
∑ Bj
– C.F.
gr
(544.5) 2 + (836.7) 2 + ... + (542.1) 2
= – 122298.83
4 x4
= 6138.35
Date of transplanting
Green manure d1 d2 d3 d4 Gi
g1 104.4 181.1 175.0 107.7 568.2
g2 124.3 190.7 205.8 144.4 665.2
g3 153.3 216.1 223.4 130.6 723.4
g4 162.5 248.8 270.2 159.4 840.9
Dj 544.5 836.7 874.4 542.1 2797.7
116
Biometry
Total SSGxD
Total SSGD = – C.F.
r
(104.4) 2 + (181.1) 2 + ... + (159.4) 2
= – 122298.83
4
= 8908.91
MS E ( a )
c.v. for green manure = ( ) 100
GM
6.32
=( ) 100
43.71
= 5.8%
MS E ( b )
c.v. for transplanting = ( ) 100
GM
8.70
=( ) 100
43.71
117
Biometry
= 6.7%
MS E ( c )
c.v. for interaction =( ) 100
GM
12.02
=( ) 100
43.71
= 7.9%
16. Standard error of the mean, difference and critical difference
a. Green manure
MS E ( a )
SE(m) = ±
dr
6.32
=
4 x4
= ±0.63
2 MS E ( a )
SE (d) = ±
dr
2 x6.32
=
4 x4
= ±0.89
CD(LSD) = SE(d) x t0.05 at error d.f.
= 0.89 x 2.262
= 2.01
b. Date of transplanting
MS E ( b )
SE(m) = ±
gr
8.7
=
4 x4
= ±0.74
2 MS E ( b )
SE (d) = ±
gr
2 x8.7
=
4 x4
= ±1.04
CD(LSD) = SE(d) x t0.05 at error d.f.
= 1.04 x 2.262
= 2.35
c. Interaction (any two means)
118
Biometry
MS E ( c )
SE(m) = ±
r
12.02
=
4
= ±1.73
2 MS E ( c )
SE (d) = ±
r
2x6.32
=
4
= ±2.45
CD(LSD) = SE(d) x t0.05 at error d.f.
= 2.45 x 2.052
= 5.03
Table 5.27. Mean values of green manure, date of transplanting and combinations
Treatment Mean
1. Green manure, G
g1 35.51
g2 41.58
g3 45.21
g4 52.56
SE(m) = 0.63
c.v. = 5.8%
2. Date of transplanting, D
d1 33.88
d2 34.03
d3 52.29
d4 54.65
SE(m) = 0.74
c.v. = 6.7%
3. Interaction, G x D
g1d1 26.10
g1d4 26.93
g2d1 31.08
g3d4 32.65
g2d4 36.10
119
Biometry
g3d1 38.33
g4d4 39.85
g4d1 40.63
g1d3 43.75
g1d2 45.28
g2d2 47.68
g2d3 51.45
g3d2 54.03
g3d3 55.85
g4d2 62.20
g4d3 67.55
SE(m) = 1.73
c.v = 7.9%
Mean separation
1. Green manure
g1
g2
g3
g4
2. Date of transplanting, D
d4
d1
d2
d3
3. Interaction, G x D
g1d1
g1d4
g2d1
g3d4
g2d4
g3d1
g4d4
g4d1
g1d3
g1d2
g2d2
g2d3
g3d2
g3d3
120
Biometry
g4d2
g4d3
Interpretation
The analysis of variance for split-block design (Table 5.26) showed that the effects
due green manure, date of transplanting and their interaction (green manure x date
of transplanting) were significant.
Among the green manure levels, g4 (52 units/plot) had the highest yield per plot,
followed by g3 while g1 had the lowest yield. Among the date of transplanting, d3 (54
units/plot) had the highest yield per plot, followed by d2 while d4 (which was not
statistically different from d1) had the lowest yield. The least yield might be due to late
planting that could not allow sufficient time for decomposition or the crop used as
green manure was inefficient to undergo decomposition and release the necessary
nutrients for the main crop (tomato) to be grown.
Of the over all combinations, g4d3 (67.55 units/plot) revealed the highest performance
may be due to the legume used could easily undergo decomposition to supply the
necessary nutrients and the date of transplanting could be the optimum that
maximized the utilization of growth resources.
5. FACTORIAL EXPERIMENTS
6.1. Introduction
There are many cases where experiments involve the study of the effects of two or
more factors. If interaction exist and are common, experiments are planned in such a
way that they can be measured and tested. This cannot be complete if we vary a
factor at a time. It is necessary to include combination of two or more factors at
different levels in the same experiment to deal with the interaction.
In factorial experiment, all possible combinations of the levels of the factors are
investigated. A factor is a kind of treatment, and in factorial experiments, any factor
will supply several treatments. The term levels refer to the several treatments within
any factor. Thus, a factorial experiment is one in which the set of treatments consists
of all possible combinations of the levels of several factors. For example, if there are
a levels of factor A and b levels of factor B, each replication contains all ab treatment
combinations. If the ab treatment combinations are replicated r times then, there will
be abr experimental units.
The term factorial refers to the nature of the treatments to be included in the
experiment but not to the experimental design. In short Factorial experiment is the
arrangement of treatments, it is not a type of design. Factorial experiments are used
in so many fields of research. They are of great value in exploratory research where
little is known concerning the optimum levels of the factors.
121
Biometry
Factorial experiments have got several advantages. When the factors are
independent, the factorial approach can result in a saving of time and material as the
main effects can describe the results and serve as hidden replication, one factor
becomes an addition replicate of the other. When factors are not independent, the
factorial approach provides a systematic set of treatment combinations for estimating
and testing interactions.
The primary disadvantages of factorial experiments are that as the number of factors
increases, the total number of treatments becomes very large and large factorials can
be difficult to interpret, especially when interactions are all significant.
6.2.2. Layout
Let us assume that there are four levels of nitrogen (n1, n2, n3 and n4) and 4 levels of
phosphorus (p1, p2, p3 and p4). Then, there be 4 x 4 = 16 combinations (treatments).
If these treatments are replicated three times then there will be 16 x 3 = 48
experimental units (plots). Let us further assume that a rice variety will be used and
data are collected on grain yield to identify the best combination or levels. As the
experiment is conducted in the field, RCBD is used. In conducting this experiment, all
possible combinations should be done (n1p1, n1p2, n1p3, n1p4, n2p1, n2p2, n2p3, n2p4,
n3p1, n3p2, n3p3, n3p4, n4p1, n4p2, n4p3 and n4p4). Secondly, these 16 treatments will be
randomized in each block independently and finally, the layout will be prepared as in
RCBD.
6.2.3. Analysis
The analysis of variance is also done as in RCBD taking combinations as treatments.
If the treatment effect is significant, then the treatment effect will be partitioned into
due to nitrogen, phosphorus and the interaction of nitrogen by phosphorus. If the
interaction is significant then, the best combination will be chosen, and if it is not
significant, recommendation will be based on the best level of each factor. The detail
analysis of variance for two factors are given below.
122
Biometry
Analysis of variance for factorial experiment is done twice. In the first part, only three
source of variations (block, treatment and error) are considered as in RCBD. If the
treatment effect is found to be significant, then the second part of analysis will be
done. Here, treatment refers to the different combinations of the levels of the two
factors.
Table 6.1. Yield data of three varieties tested at different levels of nitrogen fertilizer.
Part-I
1. The linear additive model (L.A.M.): Xikj = µ + Ai + Bj + (AB)ij + Rk + Eijk, where, Xijk =
the ith level of factor A, jth level of factor B, in kth block, µ = the overall mean, Ai =
effect of factor A, Bj = effect of factor B, (AB)ij, interaction effect, Rk , effect of block
and Eijk, = the error term associated with the ith level of factor A, jth level of factor
B, and (AB)ij in kth block.
Assumptions:
123
Biometry
SSt =
∑ Ti
– C.F.
r
(9.6) 2 + (14.29) 2 + ... + (18.18) 2
= – 1156.82
4
= 35.79
5. Sum of squares due to block (SSB)
2
SSB =
∑ Bj
– C.F.
t
(76.98) 2 + (73.97) 2 + ... + (77.21) 2
= – 1156.82
15
= 0.44
6. Sum of squares due to error (SSE) = Total SS – SSt - SSB
= 40.26 – 35.79 – 0.44
= 4.03
At this point we need to construct ANOVA and test the treatment effect as in Table 6.2
Since the treatment effect was significant, it is necessary to proceed to compute the
three components of the treatment sum of squares (variety, nitrogen and Vx N) as
follows.
124
Biometry
Part-II: Make two way table between factor 1 and 2 (variety and nitrogen) (Table 6.3)
SSV =
∑ Vi
– C.F.
nr
(72.78) 2 + (76.70) 2 + (78.68) 2
= – 1156.82
5 x3
= 1.20
SSN =
∑ Nj
– C.F.
vr
(32.02) 2 + (43.26) 2 + ... + (52.75) 2
= – 1156.82
3 x3
= 32.46
We partitioned the treatments sum of squares into three (SSV, SSN and SSVxN). We
can import the remaining two components (SSB and SSE) from Part-I ANOVA table
and then summarize the ANOVA as in Table 6.4.
125
Biometry
Table6.5. Mean grain yield (t/ha) of three varieties evaluated under 5 levels of
nitrogen fertilizer.
Treatment Mean
1. Variety
V1 4.85
V2 5.11
V3 5.25
SE(m) = 0.10
126
Biometry
CD = 0.29
2. Nitrogen
N0 3.56
N1 4.81
N2 5.30
N3 5.83
N4 5.86
SE(m) = 0.13
CD = 0.37
c.v. = 7.5%
V1 N0
V2 N1
V3 N2
N3
N4
Interpretation
The effects due to variety and nitrogen were significant while their interaction was not
implying that these three varieties were not responsive to the different levels of
nitrogen in different ways, may be due to their similar genetic background. Hence,
instead of selecting the best combination, the two factors can be recommended
independently, that is, the best variety and the most economical nitrogen dose (Table
6.5).
Of the three varieties, V3 had the highest grain yield but V3 and V2 were not
statistically different. There was a significant difference between V1 and V3 and hence
it is better to use V3.
The highest yield was obtained at N4 level of nitrogen fertilizer but N4 and N3 were
not statistically different. Hence, N3 is expected to be more economical.
127
Biometry
The difficulty of making formal significant tests using data from fractional factorial
experiments lies in the determination of proper error terms. Unless there are data
available from prior experiments, the error must come from a pooling of contrasts
representing effects that are presumed to be negligible.
The appropriate statistical analysis for data from series of experiments can vary with
the objective of the research (Cochran and Cox, 1957). The purpose of the analysis
of combined experiments is to generate valid recommendations for users to apply the
research results within a defined target area. According to Crossa (1990), data from
multi location experiments have three major objectives: to accurately estimate and
predict yield based on limited experimental data, to determine yield stability and
pattern of response of treatments (genotypes, agronomic practices, cropping
systems, feeds, etc.) across environments and to provide reliable guidance for
selecting the best treatments (genotypes, agronomic practices, cropping systems,
feeds, etc.) for use in future years and new environments.
128
Biometry
The function of the experimental design and statistical analysis of multilocation trials
is to eliminate as much as possible of the uncontrollable contained in the data.
Gauch (1988) mentions that statistical analysis of multilocation trials may have two
different objectives. With the first, a statistical model is constructed for a data set, and
success is measured in terms of model’s ability to fit this data set. With the second,
the data from a yield trial are portioned into modeling data and validation data.
7.3. Analysis
Analysis of combined experiments is done after doing some preliminary analysis.
Separate analysis of variance should be done for each experiment followed by
testing experimental errors for homogeneity and a preliminary estimate of the
interaction of treatment with the environment/season.
When considering the combed analysis of data from several experiments, the first
requirement is to assess the homogeneity of the error variance at from each
experiment. If the errors are homogeneous, the analysis can proceed. If the errors
are heterogeneous, then the data can be transformed to produce homogeneous
variance or the locations may be separated into groups within which the variance is
homogeneous.
In the process of preliminary combined analysis, the pooled error mean square could
be used to test effects of treatment, locations and the interaction (T x E). If the T x E
effect is found to be significant, then we will have to candidate mean squares to test
the effect of treatment and location, the T x E MS or the pooled MS from error SS
and T x E SS as denominator.
129
Biometry
The data presented in Table 7.1 are the mean performance of five genotypes tested
at six environments for two years using randomized block designs with six
replications.
Table 7.1. Yield (kg/ha) performance of five genotypes tested at six environments
Individual location analysis of variance is the first step towards combined analysis of
variance. One way of combing such data is to calculate the error associated with
each mean (Me’). This can be done by importing error mean square, degree of
freedom for error and number of replications from individual location analysis of
variance as follows.
E1 E2 E3 E4 E5 E6
MSE = mean square error at each environment, r = number of replicate at each environment,
E = environment, Gi = total of each genotype, and Ej = total of each environment
MSE
Me’ = = the error associated with each mean,
r
In order to combine the data the error with which each mean is measured should be
tested for homogeneity which is one of the basic assumptions of analysis of variance.
A quick test of homogeneity of variance is provided by the ratio of the largest mean
square error to the smallest mean square error in the set (Gomez and Gomez, 1984).
The result is then compared with tabulated F-value at error degrees of freedom for
individual location. In our case,
130
Biometry
F-value with f1 = f2 = 20 is 2.94. Since 2.36 is less than 2.94, it can be concluded that
the error variances are not heterogeneous.
An alternative procedure, which is more sensitive than the ratio test, is Bartlett’s test
of homogeneity of variance (Snedecor and Cochran, 1980). This test is based on the
natural logarithm of sample variances. The procedure is as follows: let kj is error
degrees of freedom for individual location, Mej error mean square for individual
location, log is the natural logarithm to the base ten, the Bartlett’s χ2 can be
calculated as follows (Table 7.2).
R j log MS E − ∑ K j log MS E
χ2 = 2.3026 [ ] , n = number of environments
1 1 1
1 + (n − 1)(∑ −
3 kj ∑kj
453.6 − 451.8
= 2.3026 [ ]
1 1
1 + (6 − 1)(0.30 −
3 120
= 4.06
Since the calculated χ2 value (4.06) is less than the tabulated χ2-value (11.07), then
χ2 test is non-significant. Hence the data can be combined because non-significant χ2
test implies the errors associated with each mean are homogenous.
(GT ) 2 (42283) 2
1. Correction Factor C.F= = =59595069.63; g= genotype & e= environment
ge 5 x6
2. Total SS = ∑x.j2 – C.F
= [(702) 2+ (770) 2+…+ (2399) 2] – 59595069.63
= 11067063.37
3. Sum of squares due to Genotype (SSG)
131
Biometry
SSG =
∑G i
– C.F., where Gi is total of ith genotype
e
(8639) 2 + (8102) 2 + ... + (8940) 2
= - 59595069.63
6
= 88499.54
1. Sum squares due to environment (SS)
2
SSE =
∑ Ej
- C.F., where Ej is total of jth environment
g
(3737) 2 + (7232) 2 + ... + (11118) 2
= – 59595069.63
5
= 10764355.37
2. Sum squares due to G x E interaction (SSGxE)
SSGxE = TSS- SSG- SSE; where G x E is the interaction between G and E
= 11067063.37 - 88499.54 - 10764355.37
= 214208.46
3. Sum squares due to pooled error (SSE)
The genotypes in the above ANOVA table are tested using pooled error mean square
and that will tell us whether the effect of genotypes is significant or not. On the other
hand, the mean square of the genotypes can also be tested using the G x E mean
square and that will tell us whether there is consistency in the genotype performance
across the environment or not. For the genotypic effect it will be
132
Biometry
MSG 22124.89
F= = = 3.07
MSGxE 10710.42
The calculated F(3.07) is greater than the tabulated F(2.87) indicating that the effect
due to genotype was significant. For the environmental effect
MS E 2152871.07
F= = = 201.01.
MSGxE 1071042
The calculated F(201.01) is greater than the tabulated F(2.71) indicating there was
significant effect of the environment.
Me' 1012.14'
SE(m)= ± =± = ±12.99
e 6
2Me' 2x1012.14
SE(d) = ± =± =± 18.37
6 6
CD = SE(d) x t0.5 at error D.F. = 18.37 x 1.98 = 36.37
Table 7.4. Mean yield (kg/ha) of five genotypes tested at six environments
Genotype Environment
___________________________________________________
E1 E2 E3 E4 E5 E6
G1 702 1498 583 2140 1454 2262
G2 770 1266 614 2071 1194 2187
G3 833 1429 723 2115 1276 2137
G4 527 1557 718 2056 1098 2133
G5 905 1482 626 2074 1454 2399
133
Biometry
INTERPRETATION
The combined analysis showed the effect of genotypes, the environment and their
interaction were significant at 0.01 level of probability. Genotype 5 had the best mean
performance across locations (Figure 7.1). It out yielded all others at E1, E5 and E6
(Table 7.4) but G1 yielded the same as G5 at E5. G1 is the second best though G1 and
G3 were not statistically different. It performed highest at E2, E4 and as equal as G5 at
E5.
Of all the testing sites E6 is the best followed by E4 while E3 is the poorest. Significant
Genotype X Environment indicates that genotypes were not consistent in
performance across the environments (Figure 7.1.)
3000
2500
2000
Mean yeild (kg/ha)
G1
G2
1500 G3
G4
1000 G5
500
0
E1 E2 E3 E4 E5 E6
Environment
7.4. Summary
Data from several experiments help researchers estimate treatment performance
more accurately, select better production alternatives, and understand the interaction
134
Biometry
There are several criticisms towards the analysis of combined experiments especially
when data violates the basic assumptions of analysis of variance. The first criticism is
that some components of the T x E SS may be much larger than others, the
interaction variance is not constant which happens when treatment response varies
greatly from place to place while others vary little or not at all. A method of coping
with this difficulty is to divide the treatment SS into a set of orthogonal components
that will supply all the necessary information. The interaction SS is partitioned in the
same way so as to isolate the interaction of each component with places. If the
interaction variance is not still constant , it is valid to test any component with its own
interaction with places. The drawback of this method is that the degree of freedom for
the denominator of F- are reduced.
A second criticism related to the assumption that the experimental error variances
are the same in all experiments. This assumption hold if all experiments have been
conducted in the same way, with the same control over environmental variables and
experimental material of the same variability. In practice this is seldom achieved
because of the natural variability of soil, climate and crop species. Therefore, the
experimental error variances varies from place to place and when we feel that the
variation is high (> 3X the smaller) it is necessary to run Bartlett’s test of
homogeneity.
Hence, it is better to use additional statistics and available soft wares and more
attention has to be devoted to the collection, analysis, and interpretation of
biophysical variables. This can help to characterize particular genotypes and
geographical regions and therefore, better explain certain aspects of the interaction.
8. ANALYSIS OF COVARIANCE
8.1. Introduction
The analysis of covariance is concerned with two or more measured variables where
any measurable independent variable is not at the predetermined levels as in
factorial experiment (Steel and Torrie, 1980). The covariance between two random
variables is a measure of the nature of the association between the two. It makes use
of the concept of both analysis of variance and of regression, i.e., it detects the
variances and covariances of specific variables to estimate treatment effect more
accurately than the use of ANOVA alone.
135
Biometry
to a single covariate that has linear relationship with the trait of primary interest. The
importance of covariance analysis in improving precision has been emphasized by
several authors (Cochran and Cox, 1957; Steel and Torrie, 1980; Gomez and
Gomez, 1984; and Walpole et al., 2002). The core remarks of the authors can be
summarized as follows. The most important uses of covariance analysis are to:
control error (increase precision), adjust treatment means; assist in the interpretation
of data, partition the total covariance into component parts and estimate missing
data.
The use of covariance to control error is by means of regression in such a way that a
certain recognized effects that can not be controlled effectively by experimental
design. The main task here is to control the variation due to the physical conduct of
the experiment but not usually on inherent variation. Examples of variation that can
be controlled by covariance analysis are: when crop varieties with different number of
plants per plot are evaluated together for yield (number of plants per unit area and
yield are positively associated in most cases) we can use number of plants per plot
as a covariate and there by estimate the true treatment effect. Estimation of the true
treatment effect will enable us to adjust treatment means. The adjusted treatment
means in turn help in proper interpretation of data.
136
Biometry
Table 8.1. Grain yield and number of plants (stand count) of eight varieties evaluated
in field experiment
Replication
Variety
Ti Si TiSi
I II III IV V
S T S T S T S T S T
1 20 30.50 16 27.25 15 17.50 12 13.25 14 18.50 107.00 77 8239
2 20 36.25 15 21.75 15 18.50 16 16.50 15 19.75 112.75 81 9132.75
3 15 24.25 12 18.75 13 20.25 11 21.75 14 18.75 103.75 65 6743.75
4 14 18.25 20 25.25 19 26.25 14 18.25 14 16.50 104.50 81 8464.50
5 20 30.50 20 36.50 15 20.50 16 23.50 15 19.25 130.25 86 11201.5
6 20 40.50 10 22.25 12 20.75 10 21.75 8 12.25 117.50 60 7050.00
7 15 20.75 16 25.75 10 21.25 12 20.25 9 10.50 98.50 62 6107.00
8 20 48.75 15 21.75 16 20.75 11 20.50 12 22.50 134.25 74 9934.50
Bj 144 249.75 124 199.25 115 165.75 102 155.75 101 138 908.5 586 66873
BjBj 35964 24707 19032.5 15886.5 13938
Si = number of plants per plot and Ti = variety total
(GT ) 2
1. Correction factor (C.F.) = where, r= number of replication and
rt
137
Biometry
(908.5) 2
= t = number of treatment
5 x8
= 20634.31
3. Total sum of squares (Total SS)
Total SS = ∑∑Xij 2 –C.F.
=(30.5)2 + (27.25)2 + … + (22.5)2 – 20634.31
= 2160.32
4. Sum of squares due to treatment (SSt)
2
SSt =
∑ Ti
– C.F.
r
(107) 2 + (112.75) 2 + ... + (134.25) 2
= – 20634.31
5
= 234.19
5. Sum of squares due to block (SSB)
2
SSB =
∑ Bj
– C.F.
t
(249.75) 2 + (199.25) 2 + ... + (138) 2
= – 20634.31
8
= 972.03
6. Sum of squares due to error (SSE)
SSE = Total SS – SSt – SSB
= 2160.32 – 234.19 – 972.03
= 954.10
Sources of F-tabulated
variation d.f. SS MS F-cal 0.05 0.01
Blocks 4 972.03 243.01 7.13* 2.71 4.07
Varieties 7 234.19 33.46 0.98 2.36 3.36
Error 28 954.10 34.08
The different components for number of plants per plot (stand) are computed as
follows.
(GT ) 2
1. Correction factor (C.F.) = where, r= number of replication and
rt
(586) 2
= t = number of treatment
5 x8
= 8584.90
138
Biometry
SSt =
∑ Ti 2 – C.F.
r
(77) 2 + (81) 2 + ... + (74) 2
= – 8584.90
5
= 133.5
4. Sum of squares due to block (SSB)
2
SSB =
∑ Bj
– C.F.
t
(144) 2 + (124) 2 + ... + (101) 2
= – 20634.31
8
= 972.03
5. Sum of squares due to error (SSE)
SSE = Total SS – SSt – SSB
= 437.10 – 133.5 – 157.85
= 145.75
Table 8.3. Analysis of variance for number of plants per plot (stand)
Sources of F-tabulated
variation d.f. SS MS F-cal 0.05 0.01
Block 4 157.85 39.46 7.57* 2.71 4.07
Treatment 7 133.50 19.07 3.66* 2.36 3.36
Error 28 145.75 5.21
(GTS xGTT )
1. Correction factor (C.F.) =
rt
(586 x908.5)
=
5 x8
= 13309.53
2. Total sum of products (Total SP)
Total SP = ∑Sij2 x Xij – C.F.
=(30.5x20) + (27.25 16) + … + (22.5x12) – 13309.53
= 712.97
139
Biometry
SPt =
∑ Ti
– C.F.
r
(107 x77) + (112.75 x81) + ... + (134.25 x74)
= – 13309.53
5
= 65.07
4. Sum of product due to block (SPB)
2
SPB =
∑ Bj
– C.F.
t
(249.75 x144) + (199.25 x124) + ... + (138 x101)
= – 13309.53
8
= 381.47
5. Sum of product due to error (SPE)
SPE = Total SP – SPt – SPB
= 712.97 – 65.07 – 381.47
= 266.43
SPE XY
6. b(regression of y on x) =
SSE X
266.43
=
145.75
= 1.83
( SPE XY ) 2
7. SSy adjusted for x error = SSy error –
SSE X
(266.43) 2
= 954.10 –
145.75
= 467.07
8. Adjusted d.f. = (r-1)(t-1) – 1
= (5-1)(8-1) – 1
= 27
SSy − adjusted for x
9. MSE adjusted =
Error d . f . − Adjusted
467.07
=
27
= 17.30
10. Treatment + Error
d.f.(t+e) = d.f. trt + d.f. error = 7 + 28 =35
SSx(t+e) = SSt of x + SSE of x = 133.50 + 145.75 = 279.25
SPxy(t+e) = SPxy trt + SPxy error = 65.07 + 266.43 = 331.50
SSy(t+e) = SSt of y + SSE of y = 234.19 + 954.10 = 1188.29
Adjusted d.f. (trt + error) = [(t-1)(r-1)] + (t-1) – 1 = [(8-1)(5-1)] + (8-1) – 1 = 34
140
Biometry
(trt + eSPXY ) 2
Adjusted SS (trt + e) = trt + e SSy –
trt + eSS X
(331.5) 2
= 1188.29 –
279.25
= 794.76
Adjusted SS (trt + e)
Adjusted MS (trt+ e) =
Adjusted d . f .(trt + e)
794.76
=
34
= 23.38
Treatment Adjusted d.f. = Adjusted d.f. (trt + e) - Adjusted d.f. error 34 – 27 =7
Treatment adjusted SS = Adjusted SS (trt + e) – Adjusted SS error
= 794.76 – 467.07
= 327.69
Treatment adjusted MS = Treatment Adjusted SS/ Treatment adjusted d.f.
= 327.69/7
= 46.81
Treatment Adjusted MS 46.81
F-cal = = = 2.71
Adjusted MS E 17.30
Table 8.4. Analysis of covariance for grain yield data with plant stand as covariate
In order to adjust mean of the grain yield, the following components are necessary
which will be illustrated in tabular form for simplifying the computation (Table 8.5).
141
Biometry
MSt X
Effective MSE = MSE adjusted [1+ ]
SSE X
19.07
= 17.30 [1+ ]
145.75
= 19.56
Interpretation
The analysis of variance for grain yield alone (Table 8.2) showed that treatment
effects were nonsignificant. However, after covariance analysis, taking number of
plants per plot as covariate, the treatment effects were found to be significant
indicating that covariance analysis helped to reduce experimental error, that is the
true treatment effects were revealed after covariance analysis (Table 8.4). The
ANOVA table for number of plants per plot showed that there was a significant effects
of stand count (Table 8.3).
142
Biometry
Covariance analysis increased the efficiency of the experiment by 74% implying that
covariance analysis is important in improving precision. Covariance analysis is also
an important method to adjust treatment means in such a way that treatments
affected by the plant stand (number of plants per plot during harvesting) was
adjusted and showed differences in their rank.
9. CORRELATION
9.1. Introduction
Correlation is used to study the relationship between two types of measurements
made on the same individuals. Correlation should not be used without first examining
the data using a scatter diagram. The scatter diagram provides a visual impression of
the nature of relation between two variables (x and y) in a bivariate data set. In many
cases, the points appear to band around a straight line. The visual impression of the
closeness of the scatter to a linear relation can be quantified by calculating a
numerical measure, called correlation coefficient.
The value of r is always between –1 and +1. The magnitude of r indicates the
strength of a linear relation, whereas its sign indicates the direction. More specifically,
r >0 if the pattern of x and y values is a band that runs from the lower left to upper
right, r < 0 if the pattern of x and y values is a band that runs from the upper left to
lower right, r = +1 if all x and y values lie exactly on a strait line with positive slope
(perfect positive relation) and r = -1 if all x and y values lie exactly on a strait line with
negative slope (perfect negative relation). A high numerical value of r, closer to either
+1 or -1 represents a strong relationship. A value of r closer to zero means that the
linear association is very weak or nil.
The correlation coefficient is closer to zero when there is no visible pattern of relation;
that is, the y value does not change in any direction as the x values change. A value
of zero could also happen because the points band around the curve that is far from
linear (may be curvilinear).
The sample correlation coefficient quantifies the association between two numerically
valued characteristics. It is calculated from n pairs of observations on two characters
(x1,y1), (x2,y2), . . . , (xn, yn). The correlation coefficient is best interpreted in terms of
the standardized observations:
Observation − means xi − x ∑ (x i − x )2
= , where = sx = .
s tan dard deviation sx n −1
143
Biometry
The sample correlation coefficient is the sum of the products of the standardized x
observation times the standardized y observations, divided by n-1.
1 x −x y −y
r= ( i )( i )
n −1 ss sy
When the pair ((xi,yi) has both components above their sample means or both below
their sample means, the product of the standardized observations will be positive;
otherwise it will be negative. Consequently, if most pairs have both components
simultaneously above, or simultaneously below, their means, r will be positive. An
alternative formula for r is used for calculation. It is obtained by canceling the
common term n-1. Then
s xy
r= where sxy = (xi – mean)(yi – mean) = covariance of x and y
s xx s yy
Total 8 16 0 0 14 10 11
x =2 x =4 sxx syy sxy
s xy
r =
s xx s yy
11
=
14 x10
= 0.93
A third option to compute r is shown below for the data in Table 9.2.
144
Biometry
(∑ x) 2 (∑ y ) 2
∑ xy − ∑ n ∑
( x)( y )
sxx = ∑ x2 − n
, syy = ∑ y2 − n
and sxy =
∑ xy − ∑ n ∑
( x)( y )
r = ___________________________
(∑ x −
∑ ( x) )(∑ y − ∑ ( y)
2
2
2
2
)
n n
8 x16
43 −
= 4
82 162
(30 − )(74 − )
4 4
= 0.93
145
Biometry
10. REGRESSION
10.1. Introduction
Regression analysis is a statistical technique for investigating and modeling the
relationship between variables (Montgomery et al., 2001). Regression is similar to
correlation as it is used for testing a linear relationship between two types of
measurements made on the same individuals. However, regression goes further in
that it is also possible to produce an equation describing the line of best fit through
the points on the graph. When using regression analysis, unlike in correlation, the
two variables have different roles. Regression is used when the value of one of the
variables is considered to be dependent on the other, or at least reliably predicted
from the other. The dependent variable is denoted by y and the independent variable
by x. Hence, the consideration is very important because regression of y on x is not
the same as regression of x on y.
Regression analysis concerns the study of the relationships between variables with
the objective of identifying, estimating and validating the relationship. The objective of
regression analysis is the development of a statistical model that can predict the
values of a variable based upon the values of another variable. In this Section the
subject will be presented with specific reference to the straight-line model. Then, on
the basis of the model, it is possible to test whether one variable actually influences
the other or not.
Regression analysis is one of the most widely used techniques for analyzing
multifactor data. Its usefulness results from the logical process of using an equation
to express the relationship between a variable of interest and a set of related
predictor variables. For example, an experimental study of the relation between two
variables is often motivated by a need to predict one variable from the other. The
director of a job-training programme may wish to study the relation between the
duration of training and the score of the trainee on subsequent skill test. A forester
may wish to estimate the timber volume of a tree from the measurement of the trunk
diameter a few meters above the ground (breast height). An agronomist may be
interested in predicting the grain yield of maize at different levels of nitrogen fertilizer.
A medical technologist may be interested in predicting the blood alcohol
measurement from the read-out of newly devised breath analyzer.
In such context as these, the predictor or input variable is denoted by x, and the
response or output variable is labeled y. The objective is to find the nature of
relation between x and y from experimental data and use the relation to predict the
response variable y from the input variable x. The first step in such a study is to plot
and examine the scatter diagram. If a linear relation emerges, the calculation of the
146
Biometry
numerical value of r will confirm the strength of the linear relation. Its value indicates
how effectively y can be predicted from x by fitting a straight line to the data.
For any experiment, n is used to denote the sample size or number of runs of the
experiment. Each run gives a pair of observations (x, y) in which x is the fixed setting
of the independent variable and y denotes the corresponding response as shown in
Table 10.1.
It is advisable to begin the analysis by plotting the data because the eye can easily
detect patterns along a line curve. The scatter diagram for the observations reveals
that the relationship is approximately linear in nature, that is, the points seem to
cluster around a straight line. Because a linear relation is the simplest relation to
handle mathematically, details of the statistical regression analysis will be presented.
A line is determined by two constants: its height above the origin (intercept) and the
amount that y increases whenever x is increased by one unit (slope). The equation of
the line fitted by method of least squares can be shown as follows.
If the relation between x and y is exactly a straight line, then the variables are
connected by the formula y = β0 + β1x, where β0 indicates the intercept of the line
with the y-axis and β1 represents the slope of the line, or the changes in y per unit
change in x.
Statistical ideas must be introduced into the study of relation when the points in a
scatter diagram do not lie perfectly on a line. The data on an underlined linear
relation that is being masked by experimental error due to in part differences in
severity of the disease, application of treatment, the plot to plot variation and so on.
147
Biometry
With these points in mind, the following regression model can be presented as a
tentative representation of the mode of relationship between y and x. Assuming that
the response y is a random variable that is related to the input variable x by
yi = β0 + β1xi + ei, i = 1, …, n
where yi denotes the response corresponding to the ith experimental run in which the
input variable x is set at the value xi, e1, …, en are the unknown error components
that are super imposed on the true linear relation. These are unobservable random
variables, which we assume are independently and normally distributed with mean
zero and unknown standard deviation σ, and the parameters β0 and β1, which
together locate the straight line, are unknown.
10.3
10.3. Numerical Example
A chemist wishes to study the relation between the drying time of a paint and a
concentration of a base solvent that facilitates a smooth application. The data of
concentration setting x and the observed drying times y are recorded in the first two
columns (Table 10.2). Plot the data, calculate r and determine the fitted line.
Table 10.2. Data on concentration (x) and drying time (y) in minutes
Concentration, x Drying time, y x2 y2 xy
0 1 0 1 0
1 5 1 25 5
2 3 4 9 6
3 9 9 81 27
4 7 16 49 28
Total 10 25 30 165 66
To calculate r and determine the equation of the fitted line, the basic quantities mean
of x, mean of y, variance of x variance of y and covariance of xy should be calculated
first using the total in Table 10.2.
10
Mean of x =
5
=2
148
Biometry
25
Mean of y =
5
=5
102
Sxx = 30 -
5
= 10
252
Syy = 165 –
5
= 40
25
Sxy = 66 – 10 x
5
= 16,
16
Then, r =
40 x10
= 0.8
16
β1 =
10
= 1.6
β0 = 5 - (1.6)2
= 1.8, then the equation of the fitted line is
ŷ = 1.8 + 1.6x .
If the sample correlation is positive, then the slope of the lease squares line is
positive. Otherwise both are negative or both zero. The main idea of regression in the
context of specific experiment can explained using the following example of a new
fungicide against purple blotch (fungal disease of onion).
An experiment was conducted to study the effect of different doses of the fungicide
treatment on the disease (purple blotch). A replicated experiment was conducted
having the different doses. Seven different dosages were used in the experiment and
some of these were repeated for more than one plot which was done purposely to
149
Biometry
show that regression and correlation can be done either on plot basis, on treatment
mean basis or a combination of the two. Each plot received a specific dosage of the
fungicide and yield data per plot were recorded as in Table 10.3.
Table 10.3. Dosage x (mg) and bulb yield per plot y (kg) of onion
Dosage, x Bulb yield, y
3 9
3 5
4 12
5 9
6 14
6 16
7 22
8 18
8 24
9 22
A glance at the table and Fig. 10.1 shows that y generally increases with x, but it is
difficult to say much more about the form of the relation simply by looking at this
tabular data that will lead us to the method of least squares.
30
25
Bulb yield
20
15 Series1
10
5
0
0 5 10
Fungicide dosage
150
Biometry
To transform the data, it is important to start with the measured values (x) and
perform transformation to get new set of values (x’). The analysis of variance or
regression is then carried out on this new sets of values. Some of the most
commonly used transformations are described below.
Log transformation is the most commonly used method. It is usually used for data
that violet normality (positively skewed distribution). The new value,
x’ = log10(x+1).
If the measurement has zero value, it is necessary to add one to all values before
applying the log transformation for the reason that natural logarithm of zero is
undefined. Adding one to each value does not affect the outcome (variance) but it is
necessary to deduct the same value at the end of the analysis from the mean.
The square root transformation is given by x’ = √x. This kind of transformation is used
when there are data from Poisson distribution (data from counts).
Angular or arcsine transformation is often used when each value in the data set is a
proportion or percentage of something. It only applies where the minimum and
maximum values are 0 and 1 (or 0 and 100%). If the original data is expressed as
percentages, then x must be converted to values between 0 and 1 before
transforming (eg. 60% = 0.6). Then,
x’ = sin-1√x or x’ = arcsin√x.
After the transformed values are generated, it is necessary to check the new values
meet the assumptions by graphing for normality, for example. If the transformed
values do not meet the assumptions, it is better to go back to the original data and try
another type of transformation. If none of the above transformations do not work, it is
better to try another simple formula like cube root transformation x’ = 3√x, 1/x or 1/x2.
It is not always possible to achieve a normal distribution using simple transformation,
151
Biometry
especially when the data have more than one peak (multi-modal distribution). This
probably indicates that the population is actually made up of two or more populations
which would be better treated as separate populations in the analysis.
152
Biometry
12.
12. REFERENCES
Cochran, W.G and G.M. Cox. 1957. Experimental designs. 2nd edition. Wiley,
New York.
Gomez, A. and A. Gomez. 1984. Statistical procedure for agricultural research. John
Wiley & Sons, Inc., New York.
Little, T.M. and F.J. Hills. 1978. Agricultural experimentation: design and analysis.
John Wiley & Sons, Inc., New York.
Mandefro Nigussie. 2005. Statistical procedures for designed experiments. Ethiopian
Agricultural Research Organization (EARO), Addis Ababa, Ethiopia. pp 241.
Steel, R.G.D. and J.H. Torrie.1980. Principles and procedures of statistics, 2nd
edition. McGraw-Hill Manual Company, New York.
Wardlaw, A.C. 2000. Practical statistics for experimental biologist. John Wiley and
Sons, LTD, New York.
153