EIAR Manual
EIAR Manual
Introduction
Field experimentation is used to obtain new information or to improve the results of
previous findings (such as new variety or improved practices). It helps to answer
questions such as:
1
objective.
In the process of experimentation there are several sources of errors that may be
encountered at all stages of the work.
• inaccurate equipment
• faulty objectives
• personal bias
• Inadequate controls /checks
• inadequate replication
• Lack of uniformity in soil fertility
• Topography or drainage
• Damage by rodents, birds, insects & diseases
• Others (poor seed germination, incorrect measurement, wrong plot size,
lack of uniformity cultural practices, competition effect ...)
In real life situation there are several factors that may affect the performance of varieties,
for example. That is because there are heterogeneity in soil, differences in slope, fertility
gradient, moisture content, etc. and it is not possible to find a plot of land that is
perfectly homogenous even for smallest possible plots. Biological world is full of
variability. This makes it difficult to visually observe and rate the performance of
variety. Therefore statistical methods are required to reveal the inherent variabilities and
Separately estimate performances of varieties that occurred due to variability and
performances due to true genetic capacity for ease of comparisons or estimation of mean
values. Statistical methods are thus used to measure the random variation (residual). It
helps to differentiate treatment variation from random variation. For statistics to
contribute effectively, the experiment must be conducted under optimum condition .
An error that might be anticipated in particular stage of a trial often propagates to all
stages that follows. This means that even if some one takes extra care in selection of
material, set up of the field, etc. a minor error that might have been contracted at one of
these stages may detract from achieving the required goal. Therefore, to be able to
achieve trustworthy conclusions and inferences, statistical methods should be considered
in designing, data collection, analysis, interpretation and result presentation stages.
Once the data is collected appropriate analytical tools will be selected and the data
subjected to analysis. This step is very crucial in the sense that it is an indication of the
fact that the statistical procedures are serially connected. That means, to produce an
acceptable result the trials must be designed properly, data must be collected properly
and correct analytical method must be used. It is quite difficult to compromise at analysis
stage if the design was initially wrong.
The Puropose of this manual is therefore to prepare a framework for linking the
designing, data management and analysis stages. In each chapter, designs or models are
2
discussed first, then programs for SAS, Genstat or MSTATc presented. Finally, outputs
from statistical packages are discussed.
This manual is prepared in 12 chapters. The first 4 chapters deals with conceptual
framework and method of analysis for some standard designs. Chapter five deals with
design and analysis of incomplete blocks. The remaining chapters entertain different
issues and topics such as specialized analysis methods, correlation & regression, cluster
analysis and comparative discussion on different statistical packages.
We are very grateful to EARO management for making available resources for
preparation of the manual. We are very thankful to research staff who read the draft of
this document and rendered constructive comments. Finally, we thank all those who
directly or indirectly contributed to the success of this work.
3
2. Developing Field Experiments
As we discussed in previous section, it is important to mention the use of statistical/
biometrical methods in relation to the stages of experimentation, because the beginning
of statistical/ biometrical planning should be right from inception of the study. Unless
problem is clearly stated, and methods to overcome the problem through experimentation
is well established, plan regarding statistical component will not be successful.
After literature search, work plan must be stated concisely and clearly. This
constitutes a "proposal". A proposal should contain the following major points.
Introduction \ Background
Justification
Objective
Experimental procedures
Expected output, Activities and Indicator of Performance
Literature review
Materials & methods
Beneficiaries
Work plan
Financial requirement
Example of objective
4
2.2 Hypothesis Testing
Once the experiment is setup properly and the objectives are indicated clearly, the next
step is to formulate hypothesis to be tested. Hypotheses are normally derived from
objectives and they are put in the form of statement. For example, in comparing
improved variety with a standard check, the hypothesis to be tested is that “the yielding
ability of improved variety is same as that of the check” and this is normally referred to
as the null hypothesis (Ho). The reverse is know as alternate hypothesis (H1) Often,
several hypotheses could be derived from a single objective. The concept may be
elaborated as follows.
There are three possibilities for testing hypothesis. The first possibility is by assuming
that the mean yield for improved variety is calculated from sample while check mean is
an absolute mean (population mean). The second possibility is to consider both new
variety and check mean to have obtained from sample observation (which is often the
case in experimentation). The third possibility is to compare two varieties from sample
observation, often known as independent samples test.
Z = 8- µ t = 8 -8o F = Treatment MS
σ S Error MS
For the third possibility both t and F test statistics could serve, but in cases where more
than two samples are compared F test is used as an over all test to detect if any two
means are different from each after, while the t-test is used to make pairwise
comparisons. Note that for two sample cases F= t2
Example
Suppose that mean yield of standard variety is 20 per acre, with standard error (s.e) of 2
and a mean yield of 25 is observed for the new variety.
Z= 25-20 = 2.5
2
Looking at the Normal probability table, a probability of obtaining, by chances, a value
of Z greater than 2.5 or less than -2.5 is 0.0724. Therefore, a probability of 0.0724 is
called a significance level for this test. We say that the difference between the newly
observed yield of variety is significantly greater than mean of the control at 0.0724. The
question now is which levels should be termed significant and which ones are not.
5
What is the Significance level?
Basically one can calculate an exact probability level at which a particular differences is
significant, but in this way there could be hundred significance levels and it may not be
possible to standardize results. Therefore, statisticians and subject matter experimenters
agreed to set threshold value of probability below which significance can be declared.
This value was chosen to be 5% but, there are numerous probability points below 5%. To
solve this problem, again, experimenters agreed to pick 5%,1% and 0.1% levels since
this levels believed to give range of significant level for most practical purposes.
1) One may demand too high a level of significance and accept the null hypothesis
Many occasions when in fact it is not true.
2) Too low level of significance will mean that the hypothesis is often rejected
when in fact it is true.
The balance between the two has to be struck by considering their likely consequences.
In practice both too high and too low level of significance have advantages or
disadvantages. Consider the following hypothesis:
1) Suppose that this hypothesis is tested at 5% level. We are saying that if a mean
difference that produce significant result at 5% is obtained then we conclude that
irrigation has an effect. But, look at it from another angle. If all possibility are
explored, however, there might be other ways of increasing yield further. But due
to the fact that we concluded significant result at 5% and recommended irrigation to
be installed could stop to look for other alternatives of increasing yield. Therefore,
opportunities for increased yield might have been lost just because we selected 5%
level.
2) If we choose 1% level, on the other hand, we might declare non- significant result
6
which could have been significant at 5%. In this case, we are overlooking the
improvement that irrigation contributed to yield increase. Therefore, we are going
to advise policy makers that irrigation is not important (whereas in reality there is
some evidence of its importance at higher probability level), consequently
expensive equipment purchased for the irrigation purpose might be dumped.
The above two statements are related to type I and type II errors.
1) Type I error :- reject Ho, when it is true. It is represented by α
2) Type II error :- accept Ho when it is false. It is represented by B.
3) Power of a test :- is probability of accepting H1, when it is true
The lesson from this example is that both high or low significance levels have their own
advantages and disadvantage and the choice of the levels must depend on the nature of
the trial. For example, tests used in social sciences normally use high probability levels
(like 10%) because the objective is often to detect small difference, and the results may
not bring an irreversible problem on the subject of study even if the result is wrong. On
the other hand, in medicine very small level are used, because more precision is required
to protect life.
7
3. Principles of Experimental Design
In this section reasons for applying statistical design will be assessed and important
concepts and terms required to implement designed experiment will be discussed in
detail.
In any experiment, i,e., a plant breeding trial, fertilizer trials, livestock trial, etc., the
design must give fairly reliable measures of relevant sources of variation.
For example, in a comparison of two varieties if one variety is planted in one plot and the
other Variety on another plot:
Experimental error
The random variation among experimental units occur due to biological variation, soil
variation, variation in technique, etc. If the breeder wants to test significance of the
difference between varieties and to compute an interval estimate of difference, one must
have a measure of experimental error.
8
Example: Experiment on effectiveness of a certain medication.
The responses variable is time of recovery. Inherent variability may result due to:
Replication
Replication can provide an estimate of the experimental error. Here the breeder repeats,
or replicates, each variety more than once in his experiment. Replication serves a number
of purposes in a designed experiment.
For variable experimental material (e.g on- farmers field ) larger replication is required
than for less variable material (e.g controlled experimental situation). Nevertheless, it is
a good idea to always safeguard the degree of freedom for error sum of square. The
quality of an experiment is normally judged by the amount of information from which
experimental error is estimated. Some people argue that the degree of freedom for error
terms for on-station field experiments must be more than 16 and some journals indirectly
follow such rules for acceptance of manuscripts.
Randomization
9
2. To ensure independence among the observations. This is necessary condition for
validity of assumption to provide significance tests and confidence intervals.
Blocking
Blocking refers to grouping of plots in to blocks of more or less uniform plots. That is,
plots within the same block are homogeneous.
Let’s us now illustrate the use of blocking, replication and randomization using a simple
(trivial) approach. How the objective, treatment contrast and hypothesis testing exercise
are well related is best illustrated as follows:
Suppose that we are working on agronomic trail of comparing the yielding ability of a
new variety A and standard variety B. The logical hypothesis is that "A" will yield more
than "B". The objective of the experiment will be to test this hypothesis. The question
now is how testing is performed.
Suppose we prepared two plots side by side and plant one plot to A and the other to B.
A B
Grain yield of the two plots will be collected and used to compare the two varieties.
Suppose plot planted to A yielded 5 q/ha and plot planted to B yielded 4.5 q/ha.
Is it possible to say that A is higher yielding than B? why?
The answer is obviously "No!". Because there are many factors that can influence the
yields of the two plots, such as differences in soil fertility, moisture, damage etc...For
this reason even if we had planted the two plots to the same variety, the yields probably
would have been different. Therefore, the difference in yields between the two plots, one
planted to A and the other to B, could be due to something other than varietal
differences.
10
Let's now introduce some concepts. Consider two plots planted to same variety A, and
two plots planted to varieties A and B.
A A A B
In any experiment the experimental error is bound to exist. That is it can't be totally
eliminated but can be minimized. A concept that should be clear here is that one can
change the size of experimental error in experiment, but one can’t change the real
difference between the two varieties. This implies that the ability to detect a true
treatment difference depends on the ability to minimize or precisely estimate
experimental error.
If two people conducted an experiment with same varieties at different place, they would
come up with different conclusions simply because of difference in their abilities to
measure and control experimental error.
The question now is how can we measure and control experimental error?
Since the experimental error is the difference between plots treated alike, the way to
measure it is to treat a number of plots alike. This is known as replication. Each
treatment must be replicated a number of times in experimental plots.
A B A B
B A B A
Rep I Rep II Rep III Rep IV
11
B A
B A
When we try to estimate variety differences, the difference may not be only real
differences but also fertility differences between plots. That is, estimates are biased in
favor of variety B. This incorporates an idea known as systematic error. It is an error of
deliberately allocating variety B to good plots.
This method, therefore, reduced experimental error by the amount of systematic error,
which is added to variety differences by misleading conclusions.
Having obtained unbiased estimate of experimental error, we now look for possibilities
of reducing experimental error. This is what we call method of error or local variation
control.
There are several techniques for this. One of them is "Blocking". We will see in next
session.
12
4. Experimental Designs for Field Experiments.
To select appropriate experimental design that best suit objectives of an experiment, the
following must be known.
In selecting appropriate design it is a good idea to go for the simplest design that controls
variability and attains the desired precision. Commonly used designs are, CRD, RCBD,
Split plot, Strip plot, Lattice.
It is a flexible design. The statistical analysis is also simple even when there are unequal
replications or missing values.
Suppose that we have 8 uniform experimental plots for a two variety comparison
1 5
2 6
3 7
4 8
Since we believed that the 8 plots are uniform and no systematic variation, the next step
is to randomly allocate the two varieties to the plots.
For a random allocation we can use either what is called random number table or lottery
method.
Random numbers are found at the back of most statistical textbooks. They are pseudo
Random numbers generated on computer using random number generator.
To allocate varieties to plot randomly select one of them, say A. Then start at any point
in the table and move in any direction you like considering only the first digit. Allocate
A to the number you draw, as long as the number you picked is less than 9.
For example, by moving downwards starting from top, we obtained numbers 4,8,2,1,
skipping repetitions and those greater than 7. Therefore, variety A will be allocated to
plots 1,2,4 and 8, and variety B to the remaining plots.
13
Method of Analysis
Taking best estimators for parameters and rearranging, summarizing and squaring gives
Ho : µ =µ2 . . . or
Ho : αi = 0 for all ;
ANOVA
If F- calculated is larger than the tabulated value then we reject the null hypothesis and
conclude that there are significant differences among the treatment means.
It is used to measure the reliability of the experiment. The theoretical framework is that
if an attempt is made to compare two experiments based on mean value then the
comparison may be biased since measurements may be done on same scale. On the
other hand if variance is used for comparison, experience showed that variance is highly
related to mean and if mean is large variance tends to be large invalidating the
comparison. Therefore, to compare two similar experiments, a ratio of variance by mean
14
is selected as this quantity has no unit of measurement and since it remains constant
throughout.
_
CV = (√MSE) x 100 expressed as a percentage , where Y = Grand mean
_
Y
Usually the common ranges of CV under controlled experiments are:
Example
Treatment (Diet)
A B C
14.20 12.85 14.15
14.30 13.65 13.90
15.00 13.40 13.65
14.60 14.20 13.60
14.55 12.75 13.20
15.15 13.35 13.20
14.60 12.50 14.05
14.55 12.80 13.80 Total
Sum y1. = 116.95 y2. =105.50 y3. =109.55 y..=332.00
_ _ _ _
Mean Y1= 14.691 Y2=13.188 y3= 13.694 Y =13.833
To get the entries of the ANOVA table first compute the correction term
ANOVA
Source d.f SS MS F
Total 23 12.2683
Diet 2 8.4277 4.2139 FT = 23.04**
Error 21 3.8407 0.1829
15
Cv = (√.1829) 100 = 3.1%
13.83
Presentation of results
Here diet is a treatment to be tested using F- ratio. Note that degree of freedom for the
error term is 21 which fairly give a good estimate of experimental error. Remember that
if not sufficient information (df) is used for estimating error, then the F – ratio may not
be valid and all tests may not make sense. Given tabulated F- values for 5% and 1%.
Therefore, the differences among the weight gain means of the diets are highly
significant.
This test is known as an overall test of difference between treatments, because so far we
could only know that the 3 diets are not the same or atleast one of them are different
from the other two. The next step is to investigate which ones are different from each
other. t – tests are often used for this purpose.
_ _
t = Y1 - Y2 = 14.7 – 13.2
√(2*S/n) √ (0.183*2/8)
= 5.32
But, t0.01(21)< 3.5 therefore, diet A is superior to B. t – value for comparing diet B with
C is also 2.4 which is significant at 5% . Now let’s ask a question. The difference
between means of B and C is 0.5 but it is significant at 5% level, why? Although the
numerical mean difference between the two is small, because of small experimental error
(0.183) even such small differences turned out to be significant. That means not only
mean value, but level of variability also contributes to significance tests. The standard
error of a mean is given by:
√0.183/8 = 0.15
Which is a good indicator of the level of precision with which the mean has been
estimated. Standard error for differences between two means is given by
The same result may be obtained by using different statistical program. For example, the
following SAS command produce ANOVA result that follows.
16
infile “filename”;
input diet weitg ;
run;
Proc ANOVA;
Class diet;
Model weitg =diet;
Run;
Unequal Replications
If the treatments are not equally replicated only minor modifications are necessary both
for hand calculation and for the SAS Program.
Proc GLM;
Class diet;
Model witg = diet;
LSMEANS diet ;
Run
Here the LSMEANS produces adjusted means for the diet.
17
4.2. Randomized Complete Block Design (RCBD)
CRD assumed a world with little variability, which is not practical in most cases,
because there are always fertility trends which casts problem on application of CRD. In
the presence of trend the use of CRD is misleading because there is a chance that one of
the varieties will receive good or all bad plots which leads to a biased estimate of error.
Suppose that we have a plot of land to test yielding ability of 4 varieties, A, B, C, D and
suppose that fertility increases as indicated by an arrow.
..........>
This means that the land is no more uniform and the use of CRD is not supported.
Next, we have to divide the land into blocks of 4 plots so that plots within block are as
uniform as possible. Once the size of land is determined it is the plot size which
determines the number of blocks in this particular example. For simplicity, let’s assumes
we require two blocks. Blocking will be perpendicular to the direction of fertility
gradient. Divide the field into equal parts, and divide each block into 4 plots in the
direction of the fertility trend.
Block I Block II
Next, randomly allocate the four varieties to the four plots in each block. Note that the
advantage of RCBD over CRD is that it separately calculate between block variation
which other wise would have entered into the experimental error.
Block I Block II
A C
C A
D D
B B
18
Analysis of variance of RCBD
X ij = µ + αi + βj + εij
= µ +(µi. - µ) +(µ.j -µ) +(X ij - µ i. -µ.j +µ)
∑∑(x ij - 8..)2 = ∑∑(8i. - 8..) 2 + ∑∑ (8.j - 8 ) 2 + ∑∑( Xij- 8i. - 8.j + 8..) 2
ANOVA
Source d.f SS MS F
Total rp-1 SST
Block r-1 SSB MSB FB = MSB/MSE
Treatment p-1 SSt MSt Ft = MSt/MSE
Error (r-1)(p-1) SSE MSE
Coefficient of variation :
cv = (√MSE )100
=
y
Significance tests:
If Ft exceeds the table value (Fα With p-1, (r-1)(p-1) d.f.) we reject the null hypothesis
and conclude that there are significant differences among the treatment means.
19
Relative Efficiency:
To see how much precision has been gained by a randomized block design in place of
the simpler, less restrictive completely randomized design, we use the relative efficiency
, RE as
RE = (r-1)MSR+r(p-1) MSE
(rp-1) MSE
If RE>1 blocking has been effective. The quantity (RE-1) measures the percentage
increase in precision due to blocking.
Example
To compare the effect of six different sources of nitrogen on the yield of wheat 24 plots
were grouped into four blocks with soil type as the blocking factor(block size = 6).
Soil Type
Treatment I II III IV Total Mean
1 32.1 35.6 41.9 35.4 145.0 36.50
2 30.1 31.5 37.1 30.8 129.5 32.38
3 25.4 27.4 33.8 31.1 117.7 29.42
4 24.1 33.0 35.6 31.4 124.1 31.02
5 26.1 31.0 33.8 31.9 122.8 30.70
6(control) 23.2 24.8 26.7 26.7 101.4 25.35
Total 161.0 183.3 208.9 187.3 740.5
Mean 26.83 30.55 34.82 31.22 30.85
20
MSB = SSB/r-1 = 192.55/3 = 64.18
MSE = SSE/(p-r)(r-1) = 44.53/15 = 2.97
Presentation of results
The difference between this and CRD ANOVA is the presence of block. If this data was
assumed as if it came from CRD, then the error sum of square would have been 237.08
since this block effect would be added on to the error.
Since Ft = 17.19> 9.72 the effect of the different sources of nitrogen on the yield of wheat
indicates highly significant differences among sources. Blocking by soil type increases
the efficiency of the randomized block design by 169% compared to a completely
randomized design. As we described earlier the ANOVA table showed an evidence of
differences between the six treatments. But we still do not know which ones are different
and need to do individual comparisons of means. At this point there are a number of
alternatives either to run multiple range tests or contrasts. This will be discussed in
section 4.4. The important points to remember in RCBD is that also we have two sources
of variation we are interested to test only the treatment and not the block effect, because
the objective is not to detect differences between blocks but control for variability.
However, if we are interested to know whether blocking is effective relative efficiency
could do. Suppose that blocking was not effective, shall we change our analysis to CRD
or report as it is. This is a source of controversy for biometricans themselves, but one
important thing to note is that the major difference between designs is randomization.
Here since in RCBD treatments were randomized within block there is a strict restriction
for treatments and this makes it different from CRD where treatments are randomized
across the experimental land. Therefore, trying to analyze an experiment designed in
RCBD with CRD will simply override this reality and affect the result.
Same analysis could also be done using computers. For example, to use MSTATC, select
“FACTOR” from the main menu and then select “RCBD 1 factor” among the 35 sub-
21
menu listed. The ANOVA structure is exactly the same as the one showed above. In
addition MSTATC gives LSD value and S.e (sy) for different mean group.
To use SAS for same analysis, the procedure may be given as (Omitting data step):
Proc ANOVA;
Class rep treat;
Model yield = rep treat;
Run;
Note that here we have two factor columns, rep & treatment and the model statements
contain both of them. Here the output is two-staged. The first part of the ANOVA
contains model & error components only, the second part splits the model component in
to two, Nitrogen & soil individual effect.
SAS output for the rcbd design
22
They are
1) Often the amount of variability is found to be related to the mean level of the yield in
an experiment. Hence, CV is more stable quantity than the unit standard deviation.
Therefore, CV helps to compare precision of two similar experiments.
2) Sometimes the experiment is stated in terms of being able to detect, with a given
probability, a stated percentage difference between treatment means. Thus CV can
be used to decide how many replicates are needed.
S = (C.V) 8
p(Z > x-8 ) = α
S/√n
⇒ Z = x-8 x-8 =d
S/√n
⇒ ( S/√n ) Z = d ⇒ n = (SZ)2
d2
where X= an observed value.
_
X= mean value
S = Unit standard error
d = mean difference
Z = the Z score under normal distribution.
Drawbacks
1) Unfortunately the amount of variation is not simply related to the mean , and CV can
often be a meaningless quantity.
Examples
In this example similar experiments were done 3 times and the CV are nearly constant.
Therefore, it is now meaningful to quote CV for future use and reference when further
experiment is preformed on same line. On the other hand, in the following experiments
23
the CV are not constant (assuming that the experiments were done under acceptable level
of precision). Hence there is no guarantee that when the next experiment is done the CV
would be close to these ones. Therefore, CV here is not useful as such.
2) As in the above example, CV can only be useful quantity, where experience shows
that it does tend to take approximately the same value in different experiments.
For example CV, of whole lactation milk yield has been found to be 25% (mead
and curnow1998). This enables us to calculate with certain amount of confidence,
the number of cows that would be required to achieve some degree of accuracy in
the measurement of lactation milk yield in the future.
3) For some measurements the s.e may be just as stable as the CV Hence CV may
not be necessary.
24
4.3. More Restrictive Designs
In randomized block design we control one source of variation by grouping the
experimental units into blocks. There are experimental designs which control two or
more sources of variation.
The Latin square design controls two sources of variation. To construct a Latin square
for P treatments we require P2 experimental units. These units are classified into P
groups, of P units each, based on one source of variation. This is commonly called the
row classification. The units are then grouped into P groups, of P units each, based on
the second source of variation. This is commonly called the column classification.
------------>Fertile
Slopy
Columns
Row
Columns
A B C
Rows B C A
C A B
Advantage of Latin square is again to reduce experimental error. This can be shown as
follows:
Total variation = treatment variation + error.
25
But error = random + row variation + column variation.
Therefore, it removes the amount of row and column variations from the error term.
Latin square design is not suitable for large number of treatments. Because since the
number of replications is equal to number of treatments, for large treatments say 25 or
more, one requires a lot of replication which may not be attained, and randomization
procedure itself is difficult.
What about RCBD ? The answer again is no. Remembering the fact that in RCBD plots
within a block are expected to be uniform, as number of plots per block increases block
tend to lose uniformity. Therefore, it is advised not to use blocks of more than about 16
plot for small- sized plots and even less for relatively large plots, except in a well
established research centers, where research plots are made relatively uniform over time.
Loss of homogeneity in a block has several consequences. The first is that the treatments
applied may not be compared efficiently within a block. As a result the difference
between any two treatment in a block will also include differences between blocks which
may lead to failure of additivity assumption and the use of ANOVA in such condition
will not be valid. One of the major problems in this regard in this country is that
optimum plot size have not been determined for most crops and optimum block size are
thus not known. Attempt has been made by biometrics services to elevate this problem.
Work has been done to determine optimum plot size for wheat and same activity is under
way for potato.
Therefore, complete block and Latin square do not serve for variety trail, since several
entries are normally handled in variety trial.
Analysis
Model : µ + αi +βi + γk + εijk
Where µ = overall mean
αi = treatment effect
β; = row effect
γk = column effect
εij = error term
Using the best estimates for the parameters of the given model and squaring and
summing over all subscripts we get,
∑∑∑ (xijk - 8...)2 = p2 ∑ (8i..- 8...)2 + p2∑(8.j. - 8...)2 + p2∑ (8..k - 8...)2
+ ∑∑∑(xijk - 8i.. - 8.j. - 8..k + 28…)2
The computational formula is
26
where CF = X2 .../p3
Source d.f SS MS F
Total p2-1 SSTo
Rows p-1 SSR MSR =SSR/ (p-1) FR = MSR/MSE
Columns p-1 SSC MSC = SSC/(p-1) FC = MSC/MSE
Treatment p-1 SSt MST = SSt/ (p-1) FT = MSt/MSE
Error (p-1)(p-2) SSE MSE = SSE/(p(r-1))
Example
Age group
Total diets
A B C D E
17.87 17.52 17.41 17.18 14.58
27
Presentation of result
There are highly significant differences among the diets. There is also evidence of
significant differences among the weight groups.
Age groups appeared to have little effect. Therefore, further experiments in this or
similar sheep might use randomized block design with weight groups as blocks, since
blocking by age group was not useful. Since there is evidence of differences between the
5 diets, the next step is to investigate which ones are different from each other by using t-
test, LSD or other multiple range test if required.
SAS procedure for analyzing the above data set is as follows, omitting data entry step.
Proc ANOVA;
Class weight age diets;
Model Y = weight age diets;
Run;
Relative Efficiency
The effectiveness of either the row or column grouping may be estimated by computing
the relative efficiency with either rows or columns as blocks in a randomized block
design.
28
For Rows : RE = MSR +(p-1) MSE
p (MSE)
Then the treatments are randomly assigned to rows and columns such that each treatment
occurs only once in each row and column.
The result shows that both rows and colums are effective but rows are more important
for blocking.
Restriction
From a practical standpoint there are some restrictions on the number of treatments
which can be tested in a Latin square
These groups of designs have come to effect as a solution to some limitation in Latin
square. For example, if P is Small then the d.f for error may be increased by using more
than one square.
multiple latin square could be applied in two ways. The squares could be taken
independently or they could be merged to form a rectangle in which each treatment
appears once in each column and twice or more in each row depending on how many
squares joined. Such difference also leads to different forms of analysis. Example given
below corresponds to the case of independent squares .
Source d.f.
Total np2-1
Squares n-1
Rows in squares n(p-1)
Columns in squares n(p- 1)
Treatment p-1
Error n(p-1)(p-2)+(n-1)(p-1)
29
Example
A researcher wanted to compare the effect of three feed additives, A,B,C, on milk
production of diary cattle. The animals had to be individually fed in digestion stalls, of
which only three were available. He decided to run a series of Latin squares. Since
three additives were to be tested, and since only three digestion stalls were available, he
decided to use 3 x 3 Latin squares with rows representing feed periods and columns
representing cows assigned to the individual digestion stalls. To obtain sufficient
degrees of freedom for error and to obtain sufficient replication for reasonable
precessions, he decided to run four squares. The experimental plan and the milk
production (pounds) during last 5 days of each 7-day period are given below.
Animal
Square Period 11 12 13
11 C 115 A 139 B 127
I 12 A 138 B 209 C 224
13 B 125 C 186 A 172
21 22 23
21 C 176 B 163 A 135
II 22 A 186 C 201 B 175
23 B 146 A 101 C 134
31 32 33
31 C 186 B 194 A 166
III 32 A 130 C 180 B 162
33 B 123 A 97 C 137
41 42 43
41 A 128 C 154 B 150
IV 42 C 137 B 129 A 106
43 B 164 A 138 C 168
ANOVA
Source d.f. SS MS F
Total 35 33,624.97
Square 3 1,738.30 579.43 6.25
Period in square 8 19,094.67 2,386.83 25.74
Cow in square 8 5,944.67 743.08 8.01
Treatment 2 5,549.05 2,774.52 29.92**
Error 14 1,298.28 92.72
30
CV = ( √ 92.73 ) 100 =6.3%
152.53
Summary of results
There were highly significant differences in mean of 5-day yields among the three
additives. Removal of period differences was most effective, while removal of differences
among squares was least effective. It is suggested that a similar design be used in future
work of this type.
Not that it is not possible to analyze data from such designs in MSTATC
SAS procedure for analyzing the above data set is as follows (omitting the data step)
proc GLM
class square period cow feed;
model yield=feed square period (square) cow (square);
run;
31
4.3.2 Cross-Over Designs
In diary husbandry and biological assay a design has been used which closely resembles
the Latin square but may have some advantages when the number of treatments is small.
The cross-over design is particularly appropriate when the difference between the rows is
substantially the same in all replicates. Even if the difference between rows is known to
be variable, the cross-over design may be preferable in small experiments where few
degrees of freedom are available for error.
In the simplest case where there are two treatments A and B, the units are first grouped
into two pairs as if a randomized block design were to be used. Suppose that from
previous knowledge one unit in each pair is expected to give a higher response than the
other and that the difference is in favor of the superior unit is expected to be about the
same in all pairs. It will clearly be advisable to insure that each treatment is applied to
the “better” members in half the replicate and to the “poorer” member in the other half.
The pairs of replicates, which must be even in number, are divided at random into two
equal sets, the first set to receive treatment A on the superior member of each pair, and
the second to receive B.
In diary husbandry the cross-over may be used to compare the effects of two feeding
rations on the amount and quality of milk produced by the cow. Since cows vary greatly
in their milk production, each ration is tested on every cow by feeding it during either the
first or the second half of the period of lactation, so that each cow gives a separate
replicate. The milk yield of a cow declines sharply from the first to the second half of its
period, so that the first half is always “better”, in the sense above. Whether a cross-over
is superior to a set of Latin squares depends on circumstances. The rate of decline is not
constant from cow to cow; it is greater in general for high yielding ability, each pair
being made a separate 2x2 Latin square. This plan likely to give a smaller error than the
cross-over, though sometimes not sufficiently smaller to counterbalance the loss of
degrees of freedom.
The design can be used with any number of treatments, subject to the restriction that the
number of replicates must be a multiple of the number of treatments. With three
treatments, for example, a plan can be drown up from the 3 cycles ABC, BCA, CAB,
where the order of the letters denotes the row to which each treatment is applied. Each
cycle is allotted at random to one-third of the replicates. For higher numbers of
treatments a design is constructed in the same way from the columns of any Latin square.
When the number of treatment exceeds four, however, the degrees of freedom for error
are sufficiently large so that a set of Latin square is usually preferable.
Once the cross over effect is taken care of by such design, then the analysis in just like
that of Latin square.
32
Columns (Replications)
1 2 ... a Total
First (poorer) AX11 BX12 . . . BX1a X1.
Second (better) BX21 AX22 . . . AX2a X2.
Totals X1. X2. ... X.a
Columns db/(a-1)
Rows ( t-1)
Treatments ( t – 1)
Error (t – 1)( r – 2)
Total (tr – 1)
4 X 4 Greco-Latin Square
Row 1 2 3 4
1 Aα Bβ Cγ Dδ
2 Bδ Aγ Dβ Cα
3 Cβ Dγ Aδ Bγ
4 Dγ Cδ Bα Aβ
33
Analysis of variance
Using the best estimators for the parameters, transposing 8 to the LHS , squaring both
sides and taking the sum over all the subscripts, be obtained as follows: the ANOVA
equation.
SS Latin = ∑ (latin)2 - CF
N
SS Row = ∑ (Row total) 2 - CF
n
SS Column = ∑ (Column total) 2 - CF
n
SS Greek = ∑ (greek) 2 - CF
n
SST = ∑∑ X2 jk - CF
SSE = SST - SS lattin -SS Row - SS column - SS Greek.
where CF = ∑ X2jk
n2
Source d.f SS MS F
Total n2-1 SST
Row n-1 SSR
Column n-1 SSC
Latin n -1 SS Lat
Greek n-1 SSGree
2
Error n - 4n +3 SSE
34
The Analysis is
{
1 if treatment i appears in block h
nih=
0 otherwise.
1
Here, note that adjustment is required for SSt because all treatments did not appear in all
rows.
Example
S1 S2 S3 S4 Total
D1 A:31 B:24 C:20 D: 20 95
D2 B:21 C:27 D:23 E: 25 96
D3 C:22 D:27 E: 25 A:29 103
D4 D:20 E:25 A:33 B:25 103
D5 E:18 A:37 B:24 C:24 103
Total 112 140 125 123 500
Feed A B C D E
Total 130 94 93 90 93
35
Analyzing the data as follows
ANOVA table
Source d. f S.S MS F
Feed (adj) 4 283 70.75 12.75**
Days 4 17 4.2
Stall 3 79.6 26.5
Error 8 44.4 5.55
Total 19 422
Presentation of results
Due to incomplete block, i.e each treatment could not appear in all rows, feed sum of
square was adjusted for days and stall. We will consider the concept of incomplete
block design in the next chapter in detail. Also note that we are interested to test effect
of feed only and the other two factors were used as a local control.
36
The SAS procedure for analyzing the above problem is as follows:
data lines;
do days=1 to 5;
do stall=1 to 4;
input feed$ yield @@;
output;
end;
end;
datalines; /* this replaced cards commands in new version*/
a 31 b 24 c 20 d 20
b 21 c 27 d 23 e 25
c 22 d 27 e 25 a 29
d 20 e 25 a 33 b 25
e 18 a 37 b 24 c 24
;
run;
proc GLM;
class days stall feed;
model yield = days stall feed;
run;
Error 8 43.90000000
37
There are two major approaches to mean comparison, multiple range tests and contrast
methods.
Multiple range (comparison) test is the process of - making all possible pair wise
comparisons of all treatment means. For example experiment with 5 treatments have (52)
=10 pairwise comparisons.
When do we use
• Do not use with structured treatments. That is multiple comparison tests should not
be carried out on variables with varying levels, for example 0,50,100,150 levels of
fertilizer.
• Desirable to make pre- determined comparisons
• Appropriate when individual comparison rather than the whole experiment is the
conceptual unit of interest.
• Making comparisons suggested by data, or making all possible comparisons not
initially planned invites for errors known as comparisonwise, experimentwise or
familywise.
Note :- If one is prepared to test unplanned comparisons by using multiple range test,
he/she has to take the risk of getting larger error rates and should be cautious about the
result. There are basically three types of errors.
Comparisonwise error can occure in planned experiments as well, but that of non-
planned experiments are much greater.
If, for example, we have five treatments in an experiment and interested to do multiple
comparisons, probability of obtaining at least one significant difference is equal to 1-(1-
38
0.05) 10 = 0.40. This means we have very high probability of obtaining atleast one
significant difference when there is in reality none. As we also see from the formula, the
probability increase as the number of comparison increases. That is why the use of
multiple range test is criticized.
NOTE :- In an experiment where only a single comparison is made, the two error rates
are the same.
When using protected LSD (i.e. we first used F test ) the unit for judging the error is no
more comparisonwise , but experimentwise.
There are a number of multiple range tests in practice. Let’s now look at them one by
one.
The procedure is to compare any pairwise mean difference against the calculated LSD
value and conclude there is difference between the means if the difference between two
mean is greater than LSD value. The advantage of LSD over other multiple range test is
that;
• It is simple
• Uses pooled variance or error mean square from ANOVA table.
Steps
∗ Put the treatment averages in ascending order (Rank them)
∗ Compute the largest mean and compare smallest with the highest.
∗ Then the next smallest with the largest, so on and so fourth,
∗ Continue this way until all are compared
39
For hand calculation of Duncans multiple range test consult Gomez and Goemez. For
computing purposes MSTAT and SAS gives multiple range test. But, packages like
GENSTAT do not because they want to discourage the use of multiple range tests.
Drawbacks of Duncan’s
b) It needs multi-valued critical value so that the difference between treatments required
for significance depends on the number of treatments in experiment. In reality
however, it is hard to believe that the difference between two treatment means
depends in any way on what others were included in the experiment.
Drawbacks
• Conservative
• Low power
• Multivalued critical value
4) Turkey's test
Tα =qα (a,f)Syi
Drawbacks
• Needs in reduction of Type I error which is very theoretical
5) Scheffe's Test
Draw backs
40
Example
Suppose there are three researchers A, B and C. And further suppose that they all want
to compare 10 varieties of wheat in 4 blocks of 8 plots.
Researcher A :-
His plan is to look for a variety that give highest yield among the 10 varieties. Hence he
planned to compare all possible pairs of varieties. After harvest he used 5% sig level and
carried out LSD to compare all 45 pairs.
1) Since his conceptual unit of interest is individual pairs of comparisons, and not the
entire experiment (the family of all 45 pairs), he have not committed
45
expermentiwise error rate(which is = 1 - (0.95 ) ) (cramer & waker)
2) But since the LSD is not protected, he expects to make about (0.05× 45= 2.25) 2
type I error out of the 45 comparisons if all 10 varieties are genetically alike.
Researcher B:-
He has initially planned to compare only three varieties he suspected could give high
yield. But after data collection, he wanted to change his mind. And wanted to test effects
suggested by the data. Thus want to compare the two extremes at 5% level (in fact after
performing F-test).
2) The comparisonwise error is much more than 0.05 , which is equal to experiment
wise error rate(steel and torie).
Researcher C :-
1) Multiple comparison procedures accept only a single MSE , but to separate mean of
interactions the split-plot design need involvement of both main-plot and sub-plot
MSE. Hence leads to erroneous conclusion.
2) Comparisons between different factors, which are not needed, may be carried out.
3) Multiple comparisons are not suitable for factorial structured treatments.
41
Conclusion
• In general comparisonwise error rate is more sound and useful than experimentwise
error rate in pairwise comparisons (cramer and walker,82). But when making
unplanned comparison (no special interest) the conceptual unit is the experiment
itself and incurs experimentwise error rate.
When in fact there is no real differences between treatment means. Therefore beware of
using multiple comparison method. Multiple range tests have a place only when the
treatments are completely unstructured , but even then one should use them keeping the
above two points in mind.
We have already discussed in detail the failure of using multiple range tests. One may
ask what is the alternative then? This will take us to another concept known as orthogonal
42
contrasts. Contrast, as the name indicate, is any comparison made between treatments.
The word “orthogonal” however qualifies it further. Two contrasts are orthogonal if the
sum of the product of their coefficients is zero. Orthogonality simply means
independence. Suppose consider a simple example where three treatments, A, B, C are
compared. In principle, since we have 3 treatments the number of orthogonal contrast
generated is only 2 (equal to treatment degree of freedom). If we compare A with B and
A with C, do we really need to compare B with C. The answer is no because by
comparing A with both B and C, we already have an idea about the comparison between
B and C. Therefore, the comparison between B and C is redundant because it is not
orthogonal (do not add new information) to the earlier two contrasts. Therefore, one must
be strategic when searching for orthogonal contrasts. It would be easy in situation where
treatments have structure. For example, if A is control and B and C are new technologies
then the contrast must be designed in such a way that treatment question are answered.
Here, the obvious approach is to a) Compare control with the average of the other two
and b) to compare the new technologies among themselves. Any third contrast would
depend on these two and can’t be included.
Consider the response of millet to added nitrogen. The following 4 treatments were used:
1. Control - no nitrogen
2. Boot - 100 kg nitrogen applied at booting
3. Thin - 100 kg nitrogen applied at thinning
4. Split - 50 kg nitrogen applied on each occasion
SAS program for calculating contrasts may be given as follows (including data step)
SAS 22:08
43
Dependent Variable: YIELD
Error 12 9.47750000
The ANOVA table shows significant treatment effect. When contrasts are considered
only boot vs. Split is significantly different from each other.
44
The analysis of covariance (ANCOVA) has the following uses
To analyze covariance by hand first obtain sum of squares for different effects, then
estimate the coefficients. We present hand calculation for ease of understanding the
concept, otherwise the steps are tedious and not convenient for hand calculation.
2. The least square estimates for µ, β, and α are Y.. , b and αi respectively
45
_
In this case the least square estimates for µ and β are Y.. and b = Txy
Txx
Therefore to test Ho : αi = 0 for all i,
Note once again that analysis of covariance by hand is tedious and it is always better to
use statistical packages.
Suppose consider a hypothetical experiment with four varieties and each variety
replicated three times. It was assumed that some of the varieties were affected by
rodents in some plots. Particularly, variety number 2, which performed very well in the
first two replications, performed poorly in the last replication. This means that some
varieties responded below their genetic potential due to factors not initially considered.
Now, the question is, what will happen if we proceed with calculation of ANOVA using
the observed data set? The answer from analysis of this data set is clear. The comparison
will be in favour of varieties not affected by rodents. Thus, the results shows non-
significance between the four varieties despite a clear mean difference for the two
replication.
This is obviously not fair and adjustment is required. This is also the difficult part of the
exercise. In animal experiments where use of covariate is relatively straightforward, and
where the purpose is to attain high final weight, the animals are expected to have
differences in their initial weight and animal with high initial weight may attain high
final weight even if the treatment was not so effective. Hence, the initial weight is
normally used as an adjustment factor for initial differences in weight and this is simply
called covariance. Here, however, we have to construct another variable to be used as an
adjustment factor. For this particular data we assumed that a three- scale rating of
condition of the yield is sufficient. Smallest scale (number 1) was given to a plot
damaged most by rodents and a scale of 3 for plots not damaged at all, while number two
was given to intermediate damage. When this variable was used as a covariate, the
result of ANOVA was completely changed, differences between treatments being
significant at about 3% probability level. The analysis has a number of benefit. The first
is that the CV for analysis without covariate was much higher than that with covariate,
possibly because the former possess unadjusted mean and large variability. Second,
variety means are adjusted for damages caused by rodent. Third, precision for ANOVA
and estimates improves tremendously. Mean is adjusted using the following formula.
46
X = Grand mean for covariate.
b represents the slope between the yield and the covariate given by
b = Exy/Exx = 20.79
The data (below) and analysis using both MSTATC and SAS is given below for
comparison.
Rep Facor A yield Covariate
1 1 30 3
1 2 78 3
1 3 23 3
1 4 28 2
2 1 45 3
2 2 80 3
2 3 36 3
2 4 65 3
3 1 21 2
3 2 18 1
3 3 34 3
3 4 29 3
The following analysis is based on the observed data without use of covariate. As
discussed earlier, differences between treatment means is not significant at 5%
probability level, when infact the difference between treatment number 2 and number 3
is visually so large.
Variable 3: yield
T A B L E O F M E A N S
1 2 3 Total
-------------------------------------------------
1 * 39.750 159.000
2 * 56.500 226.000
3 * 25.500 102.000
-------------------------------------------------
* 1 32.000 96.000
* 2 58.667 176.000
* 3 31.000 93.000
* 4 40.667 122.000
-------------------------------------------------
47
A N A L Y S I S O F V A R I A N C E T A B L E
Let’s now use covariance analysis to adjust for the damage and re-test all the effects. At
the beginning of the analysis the table of means shows unadjusted and adjusted means.
Adjusted mean for treatment two is large compared to unadjusted, while treatment
number 3 which has not been affected by the damage is adjusted downwards so that the
effect of rodent could diminish from the other treatments. This is an indication of how
important covariance analysis is in the test of hypothesis.
Variable 3: yield
T A B L E O F M E A N S
48
-------------------------------------------------
The numbers in the "Unadjusted 3" column are the means based on the
actual data in the file. The "Adjusted 3" column contains the means
adjusted with the following formula: Adj Yj = Unadj Yj - b(Xj-X)
where Adj Yj is the adjusted variable 3 mean, Unadj Yj is the
unadjusted variable 3 mean, Xj is the variable 4 mean for that
treatment combination, X is the variable 4 grand mean, and b is Exy/Exx
the slope of the data).
b = Exy/Exx = 20.79412
U N A D J U S T E D S U M S O F P R O D U C T S
K Degrees of
Value Source Freedom V4xV4 V4xV3 V3xV3
-------------------------------------------------------------------
1 Replication 2 1.167 46.083 1926.167
2 Factor A 3 0.667 -27.667 1477.583
-3 Error 6 2.833 58.917 1837.167
----------------------------------------------------------------------
Total 11 4.667 77.333 5240.917
----------------------------------------------------------------------
A N A L Y S I S O F C O V A R I A N C E T A B L E
49
Run;
Note the analysis in SAS carefully. SAS now distinguish the variable “COVAR” as
covariate (as different from factor) because this term was not declared as factor in the
class statement. Thus it is considered as a continues variable used for adjusting and
assigned 1 df.
Analysis with covariate
Error 5 612.04656863
50
Pr > |T| H0: LSMEAN(i)=LSMEAN(j)
i/j 1 2 3 4
1 . 0.0153 0.4325 0.3814
2 0.0153 . 0.0090 0.0437
3 0.4325 0.0090 . 0.1342
4 0.3814 0.0437 0.1342 .
Also note that this time SAS gave additional output. In the parameter estimate Section
of the output, slope is given for each level of the factors (rep and factor A) and the
covariate (COVAR). This is because the GLM procedure fit a Linear model of a
regression type. The coefficient given to covariate (20.794) in SAS is exactly the same
as the slope (b) calculated by MSTATC. The advantage of SAS is that it also gives the
test statistics (t= 3.16 for the covariate and P[T > t]) and respective standard error.
LSMEAN in SAS represent adjusted mean for the covariate, which is exactly the same
as adjusted mean obtained in MSTATC. In last few lines of the output, SAS gave
probability that an adjusted mean LSMEANi is different from LSMEANj. For example,
probability that LSMean1≠LSMean2 is 0.0153. Therefore, there is a significant
difference between the two. Note that if we specify the “ means” statement in SAS
program, then we can request for multiple range tests, but if we specify LSMEANS then
we can obtain probability for differences between any two treatment but not multiple
range test results.
The conventional analysis of variance usually begins from separate analysis for each
environment. Note that for separate analysis, the trials must have been designed in such
a way that sufficient information (error df) is available for valid interpretation of results.
Results of separate analysis is used for two purpose:
Once the separate analysis is done and before combining the data, it is necessary to check
for assumption of variance homogeneity. If variances of environments are not
51
homogenous, it is not possible to combine data as it is. Transformation is required to first
fulfil the assumption, and data will be combined and analysis made on the transformed
scale. A good account has been given in Gomez and Gomez regarding the types of
transformation available for each data (distribution) type and we would like to direct
readers to this book for further references.
The following data was taken from maize breading program to describe the necessary
steps required in combined analysis. The data contains 18 environments (2 years X 9
locations), each with 3 replication and 20 entries. RCBD was used for planting. Hence
the data file contain 1080 observation. The maximum MSE is 82.6 for environment 4
and the minimum is 31.7 for environment 13. The ratio of largest MSE to the smallest is
2.58. When Bartlett test was done there is an evidence of heterogeneity of variance at
about 5% level. Therefore, it was decided that a square root transformation made (after
observation of the data distribution) on yield for all environments and the resulting data
re-analyzed. A combined analysis was done in two ways:
The result of ANOVA from both types of analysis however turned out to be very similar.
This is probably for two reasons. Since the MSE ratio is around two, even if Bartlett test
gave significant result, the heterogeneity is not a such serious. Hence, the result of the
transformed data may not be very different from that of raw data. Second, finding a
perfect transformation is often difficult ( is even impossible) and it might not be possible
to stabilize variance through transformation as required.
Therefore, unless the level of heterogeneity is serious, it is better to work with the raw
data. When there is a serious variance heterogeneity, it is advisable to try different
transformation to select the one that stabilized variance best. In the Ethiopian condition
it is very common to get a variance ratio of more than two and it does not mean that there
is always problem of heterogeneity. There are several evidences that shows that Ethiopia
is variable even in small locality. That means since variance is related to mean,
performance of genotypes at different environment might be subjected to different level
of variability owing to the different soil and agro-climatic factors. Therefore we should
be able to tolerate some level of heterogeneity, possibly consider those significant at 1%
using Bartlett test as heterogeneous.
When data is combined over locations and years for analysis, the analysis of variance
structure may take different forms. There are three possibilities:
52
1. When same location and randomization is used each year
2. When same location is used each year but different randomization is adopted and
3. When different location is used each year. The structure of ANOVA is slightly
different for each of them. Data from maize was used to illustrate the issue.
The difference among the three ANOVA is mainly on the relationship between
replication on one hand, and year and location on the other hand. When same location
and randomization is used each year then replication is said to be nested under location,
but when same location is used with different randomization then replication is nested
under location by year interaction. In MSTATC this can be done easily from the
“FACTOR” menu using sub-menu number 16,17 and 18. But in packages like SAS and
GENSAT one has to specifically write the model statement so that the software
understand and differentiate the nesting condition. In SAS, the following program (old
version of SAS) produces ANOVA for same location and randomization.
For other models slight modification is required. Proc MIXED, a recent development in
SAS program, is particularly useful where split unit designs are used for multi-location
(year) experiments. The simplest example for split-plot design with 5 sub-plot and M
main plot treatments may be given below.
Proc mixed:’
Class block MS;
Model Y=M/S;
Random block M * block;
Run;
Here the “ random” statement declares an error term to be used for testing the effects.
The output for mixed procedure differs from the standard approaches very much. This is
one of the most powerful procedure recently incorporated in SAS, for more information
refer the help system in SAS or its manual.
The mixed procedure in used to fit a variety of mixed linear models. It is a generalization
of standard linear model used in the GLM procedure. The mixed procedure handles data
that exhibit correlation and non-constant variability. Therefore, it could be safe to use
mixed procedure in a situation where we do not know about the data very much. This
procedure, unlike the standard ones, model means, variances and covariances. Mixed
model is known for its two components, fixed and random effect. The name “mixed” is
meant to show the combination of the two in to one model. The fixed effects are those
whose effects are to be determined e.g variety, breed, etc.. while the random-effects are
additional unknown random variability assumed to impact the variability of the data e.g
block. In the case of fixed effect, the conclusions apply only to those factor levels which
53
were used in the study, while in the random effect case the conclusion applies to the
population from which the factor levels used in the study has been drawn from.
The mixed procedure has several options which could be used for different tests of
hypothesis about a particular data set. Effects that can used to tests other effects or that
are needed to separately estimated because of their additional randomness are included
in the “ random” statement. Whatever is not indicated in “ random “ statement must
have appeared in the model statement where fixed effects are declared.
Table 4.6.1: ANOVA for same location and randomization for each year.
Source df SS
L 8 64710.08
R(L) 18 4353.69
Y 1 12753.99
LY 8 25,088.16
RY(L) 18 2427.50
T 19 37385.76
LT 152 31127.34
YT 19 5157.52
LYT 152 25028.99
Error 684 33580.46
Total 1079 467707.50
In the case of same location but different randomization, we will have replication nested
under the interaction of location and year [R(LY)] which is basically the sum of
replication nested under location and replication nested under interaction of location by
year. The sum of square for location, year, treatment and interactions between this 3
factors, and hence tests of hypothesis, remain the same whether same location and
randomization or same location and different randomization is used. But, when
locations are different each year, the tests are also slightly different (table 6.4.2.). For
example, tests for Location by year interaction (LY), and Location by Treat (LT) are
different. Therefore, it is really important to consider seriously which one of the above
we should use to analyze multi- location/year data.
Source Df SS
Y 1 12753.99
L(Y) 16 315596.25
R(LY) 36 6781.20
T 19 37685.76
YT 19 5157.5
54
LT(Y) 304 56152.3
Error 684 33580.46
Menu item in MSTATC are not available for some multiple experiments. For example,
for a four factor experiment (say, variety, management system weeding date and
fertilizer level) in split plot design where variety and management are on main plot and,
fertilizer and weeding date are on sub-plot can’t be found on MSTATC menu directly,
but can be generated using the last sub-menu called “ other designs”. In MSTATC
design generations work on the basis of “ k- values”. The k-values are identifies of the
different effects. For example, if we have 3- factor (including rep) designated as A, B,
C, and if we need to construct ANOVA. start with K-value of 1, which gives you factor
A (which is replication), if we type number 2 it will give we Factor B. Now, if we give
K- value of 3, it will automatically take the next logical component which is A*B. But,
we do not need A*B because, we do not want interaction of rep with others. Therefore,
to Avoid A*B we ignore number 3 and type 4 which gives factor C, so on and so fourth.
Therefore a complete ANOVA may be given as:
K Source df
1 A(rep) a-1
2 B b-1
4 C c-1
6 BC
-7 error
Note that to get the error term use negative numbers. The best strategy to produce
ANOVA for design not available in the menu is to first list down the sources of
variations in ANOVA table then one can easily select the corresponding K-values. Even
if you made mistake in selection you can always go back using upwards arrow and enter
the correct k-values.
55
1. Four factor split –plot experiment in RCBD.
Here we assumed to have variety (V1) and management (V2) on main plot & fertilizer
(V3) and weeding date (V4) on split plot.
Proc mixed;
Class rep V1 V2 V3 V4;
Model Y= V3*V4/ V1*V2;
Random = rep rep * V1 rep* V2 rep * V1* V2;
run;
56
2. A three-factor split-plot design in RCBD combined over locations and year. Two
factors assigned on main plot and one factor on sub-plot.
These designs can also be analyzed using SAS by specifying which term is used to test
the intended hypothesis. Therefore, modifying the program written for split-plot design
in previous section could suffice. Equally, use of proc mixed can do the job,
57
4.6.1. Yield stability
One of the objectives of multi-location trails is to investigate an overall performance of
varieties over locations. We normally expect some genotypes to show a relatively strong
dependence on a combination of the important site characteristics while others would
show lesser dependence. The former group are known to be unstable while the latter are
stable genotypes. Therefore, to identify these genotypes a regression of genotype yield
on site characteristics is important. However, the independent environmental
characteristics which help to distinguish stable genotype from unstable one are is usually
unknown or difficult to obtain. As an alternative for the unknown environmental index,
the mean yield of all genotypes at each site is used as a measure of site performance, and
genotype yield is regressed on an overall mean yield of all genotypes in a site.
returning to the analysis of variance model for combined data where the observed mean
yield (Yij) of the ith genotype at the jth environment is expressed as:
Now since the problem lies with the interaction component (GEij), it is partitioned into a
component due to linear regression (bi) of the jth genotype on environmental mean and a
deviation (dij).
(GE)ij = bi Ej + dij
Hence
Yij = µ + Gi + Ej + (bi Ej + dij) + eij
Several authors criticized the regression approach for its statistical and bilogical
limitations. The statistical limitations are:
1. That the relative stability of two genotypes depends not only on particular sites, but
58
also on the set of other genotypes included in the analysis.
2. When only few very low or very high yielding sites are included in the analysis fit of
regression may be influenced by genotypes performance in those few extreme
environments, which might mislead conclusions.
Nevertheless, the regression approach of Eberhart and Russel for determining stable
genotype is still extensively used and reported in all journals.
Stability using Eberhart and Russel regression approach can be done in a number of
packages. MSTATC produces the slope (bi), standard errors of bi and Error mean
square, but do not produce deviation from regression as defined in literature.
When you do stability in MSTATC, you should first get mean of each genotypes for
each location, because the program do not require raw data with replications, proceed as
follows:
Yield ________
Variety ________
Treat ________
Location ________
Now from the list of variables, you can tell MSTATC on which variable number yield is
available (eg 8), etc. By now you might recognize that MSTATC requires variety and
treatment columns, but we may not have a treatment factor in our trial. Never worry
about it, you may give same variable number for both variety and treatment. If the trail
is done over locations without repeating over years then you may supply variable
number of location, otherwise create another variable called environment which is a
combination of location and year. Once you done this, the next step is to compute the
slope by giving a combination of variety and treatment. The combination must be that
variety 1 and treatment 1, etc., since both of them are coming from same variable.
For the example we considered from maize program the output is given below. The
outputs of b’s from both packages are exactly the same, but output of Agrobase has been
presented here for completeness.
59
Table 6.4.1.1: Eberhart /Russell stability regression ANOVA
Note: F-ratios computed using residual of 18.6923, with 720 df varietal mean squares
computed from (Sum of Squares)/(No. sites-2) standard error of beta = 0.1030
This is basically the beginning of investigating the G x E interaction. There are several
methods developed and available for explaining G X E and estimating adjusted yield.
This section, therefore, may just serve readers as an introduction to G x E interaction
modeling.
Interpretation of results
60
relationship exists, variety response could be predicted.
The hypothesis that there are no genetic differences among varieties for their regression
on the environment index is tested by the F-value of 1.68, which is significant in the
table above for partition of deviation from linearity, the fourth column is test for
deviation from regression for each variety. Consequently, only genotypes no 6 and 19
have deviation not significantly different from zero. The desired entiries are those which
have slope of unity, Sd2 near zero and high yield. However, those genotype which have
b close to unity and small Sd2 are know to be stable. Here most varieties have slops not
significance different from unity. But only variety 6 have Sd2 =0, thus this genotype is
the most stable. However, genotypes like 8,7,15 are also relatively stable. Genotype like
3,9 & 16 performs very well in good environments.
61
5. Incomplete Block Design
5.1. General Concept
So far we were concerned with situations where all treatments can be accommodated in a
block. That means a situation where number of treatment is equal to number of
experimental units in a block. The next question is what shall we do if we have large
number of treatments, say 20 or 30 and little resource is available which may not
accommodate such large number. This brings a new concept known as incomplete block
design.
This type of arrangement is incomplete because not all treatments appear in same block.
Each incomplete block contains three treatments.
Therefore, incomple block design may be defined as a design where all treatments can’t
appear together in a block, since number of plots per block is less than number of
treatments. Such circumstance can occur in two ways. One is when we have large
treatments and plots per block can’t accommodate them all as block may loss
homogeneity if extended. Second, practical application of treatments to blocks may not
match. That means, it may not be possible to get space for all treatments to be compared
within block.
Let’s consider two design structures and discuss the benefit and difficulties associated
with incomplete block.
(I)
1 a b b c c a b c
block 2 a a b c c a b b
3 a a b b c a b c
The design is Symmetric, which implies that all pair wise mean difference, A-B, A-C are
all Equally precise. However, this design is basically not incomplete block but repeated
appearance of a treatment in the same block. It therefore gives a contrasting idea with
that of incomplete block. The problem in incomplete block design is exactly the opposite
to this particular design. It is important to note that although all treatments did not occur
equal number of time per block, the design turned out to be in balance because of the
symmetry.
62
The following design is not symmetric, but it is in total balance, because variance of
mean differences are equal, i.e., (A-B) = var (b-c) = 2σ2
3
(II)
1 a a b c
2 a a b c
3 b c
The above design is a combination of incomplete block and extended block. The first two
blocks are extended because treatment A appeared twice in those blocks, while the third
block is incomplete as it could not accommodate treatment A. At first glance it looks as
if it can’t be balance, particularly since A have not occurred in block 3. This shows that
the idea of indirect information in treatment comparison is so powerful that sometime
mathematical approach is required to test whether a particular design is in total balance.
This concept will be elaborated further below.
The nicety of complete block is that comparison of treatments within block is possible
since all treatment appear in a block. That means treatments are compared free of block
effects. In incomplete block, however, this is not possible because some treatments may
not occur together in block. This takes us to new concepts and ideas.
Two concepts
• Intra block analysis
(Comparison of treatments within block)
• Inter block analysis
(Comparison of treatment between blocks)
This second option gives the incomplete block designs the power to compare two
treatments which have not occurred together in a block but which occurred with some of
the treatments in block which is common to both.
Example Suppose that we have five treatments to be compared in two blocks of 4 plots
each
T1 T2
T2 T3
T3 T4
T4 T5
Although T1 have not occurred with T5 to be compared in some block, they compared
through T2, T3, and T4 which occurred with both T1 & T5 in one of the two blocks. That
means, since the problem of comparing T1 and T5 is that they do not occur in same block,
and that trying to compare them will also lead to difference between the two blocks,
63
technically we say effect of T1-T5 is confounded with block effect. Therefore, we can
only compare T1 and T5 by first estimating block effect. Block effect can thus roughly
estimated by subtracting valves of T2, T3 and T4 of block 1 from that of block 2. This
difference can now be considered as block difference and used as an adjustment factor
for comparing T1 and T5. Statistical packages infact apply least square method to
estimate block effects and adjust the comparison.
The other family of incomplete block is the non-resolvable incomplete Block (IB) design.
Let’s consider the problem with an example.
Rep I Rep II
T1 T3 T5 T1 T3 T5
T2 T4 T6 T2 T4 T6
This design have two replications, but the same randomization was used in both cases and
the difficulty of analyzing the data remain the same despite additional resource use.
Hence, it is not possible to COMPARE, say, t1 - t4 since there is no inter or intra block
information available. Estimation may be improved if the randomization in the second
rep is changed, possibly with the rule that prohibit occurrence of two treatments together
twice in the experiment. Hence,T1 and T2 should be separated in the second rep.
λ (a-1) = (b-1)r
There are several types of incomplete block designs available for use to- date. It is also a
relatively difficult areas of experimental design as analysis is more complicated.
64
5.2 Lattice (quasi-factorial) designs
This design was initially introduced by Yates in 1936. They are class of BIB (PBIB).
Designs.
❶ Square Lattice
❷ Cubic Lattice
♦ Treatment should be a perfect cubic such that 3√ Treat = S (block size)
♦ It is a PBIB
❸ Rectangular Lattice
Treatments with 1st coordinate occupies P blocks, while 2nd coordinates occupy another
P blocks of size p-1.
Simple, triple and rectangular lattice designs are effective when plots are LONG and
NARROW.
Suppose we are interested to test 9 varieties but could not get a block of size 9. The
experimental land is such that width is only enough for three plots and increasing fertility
trend lengthwise. So we use balanced Lattice design with four replication
8 7 9 4 7 1 1 5 9 2 4 9
5 6 4 9 3 6 6 7 2 1 6 8
2 1 3 8 5 2 8 4 3 5 3 7
Note that each of the possible pairs of treatments appear together once, and only once, in
one of the incomplete blocks. So you should aim to obtain every treatment occurring
with every other in the whole experiment. For example, treatment 8 occurs with 7 and 9
in block 1 of rep I, with 5 and 2 in block 3 of rep II, with 4 and 3 in block 3 of rep III and
with 1 and 6 in block 2 of Rep IV.
65
which can’t be met in practice. The alternative, thus, is to switch to partially balanced
lattice design. The logic of partially balanced lattice is to use less number of replications
than that required by a balanced lattice. The obvious shortcoming is that all possible
pairwise occurrence of treatments can not be attained. This implies comparison for some
pair of treatments is more precise than others.
PROBLEM: 1.
Allocation of treatments to blocks
2.
They are not available for certain number of treatments (due to
restrictions)
∗ Square Lattices are effective when plots are SQUARE ( They control variability in
two direction)
❹ Balanced Lattices
The number of treatments must be an exact square. The size of block is the square root of
this number. Incomplete blocks are combined in groups to form separate replication. The
special feature of the balanced lattice, as distinguished from other lattices, is that every
pair of treatments occurs once in the same incomplete block. Consequently, all pairs of
treatments are compared with the same degree of freedom.
Example
For pigs of a given breed, previous experience indicated that a considerable part of the
variance in growth rate between animals can be ascribed to the litter. Hence the
experiment was planned so that litter difference would not contribute to the intra-block
error. The pigs were divided into sets of 3 litter-mates. Two sets of 3 were assigned to
each block, within a block, each set. Thus the experimental unit was composed of 2 pigs
each feeding the same pea. The plan and growth rates are given in the table 5.2 (cochran
& cox, 1957).
Rep I
Block
(1) (2) (3)
1
2.20 0.84 2.18
(4) (5) (6)
2
2.05 0.85 1.86
(7) (8) (9)
3
0.73 0.60 1.76
66
Rep II
Block
(1) (4) (7)
4
1.19 1.2 1.15
(2) (5) (8)
5
2.26 1.07 1.45
(3) (6) (9)
6
2.12 2.03 1.63
Rep III
Block
7 (1) (5) (9)
1.76 1.16 1.11
8 (2) (6) (7)
1.76 2.16 1.80
9 (3) (4) (8)
1.71 1.57 1.13
Rep IV
Block
10 (1) (5) (8)
1.77 1.57 1.43
11 (2) (4) (9)
1.5 1.6 1.42
12 (3) (5) (7)
2.04 .93 1.78
The steps in the analysis is: (The algebraic formula refers to a K x K lattice in blocks of
K units with r =(k+1) replicates).
1) Calculate block totals, replication totals, grand total G, and treatment total T.
2) For each treatment, calculate the total Bt for all blocks in which the treatment
appears. For treatment 4, this is
W = KT – (k + 1) Bt + G
67
Table 5.2.1 treatment totals and adjustment factors
4. The analysis of variance is now obtained. The total SS and SS for replication and
treatments are found in the usual way. The sum of squares for blocks within
replication adjusted for treatment effects is
∑w2 = 1.4206
k3 (k+1)
ANOVA table
General d.f d.f (k=3) S.S M.S
Replication K 3 0.0774
Treatments (k2 – 1) 8 3.2261
Blocks (adj.) (k2 - 1) 8 1.4206 0.1776
Intra-block error (k –1)(k2 –1) 16 1.2368 0.0773
Total (k3 +k2 – 1) 35 5.9609
68
Presentation of results and further discussions
In this analysis treatment sum of square was not adjusted for incomplete block, and only
the block within rep was adjusted for treatments to help obtain the RCBD error
equivalent for further comparison. The result, however, shows significant treatment
effect. Now, once the ANOVA is obtained a pairwise comparison of treatments could
also be done. Here, we are going to have two types of standard errors, one for entries in
same block and the other for entries in different blocks.
Also note that we have two types of error estimate for incomple block, intra-block error
and effective error. The former is estimated from within block entries while the later is
given above. Some packages use effective error for tests, while others use the intra-
block error.
Another important output in lattice design is efficiency of lattice design. Here, the sum
of square for blocks within replication adjusted for treatments plus intra block error
constitute the RCBD error sum of square. For this example, RCBD error sum of square
(when replication are considered as blocks) is equal to 1.4206 + 1.2368 = 2.657 with df
of 24 (8 + 16). Here, the efficiency of lattice design compared to RCBD is about 120%.
Therefore, the use of lattice was acceptable. It means that experimental plot within
replication is heterogeneous and the use of RCBD (i.e. considering replication as
complete blocks since all entries appear in each replication and ignoring the incomplete
blocks within replications) is not supported.
The adjustment factor which is, calculated in Lattice is normally an indicator of the level
of balance in the design. For example, if µ approach zero, then it means that no
adjustment is required. The adjustment factor is also useful for adjusting treatment
means, since the row entry means obtained are affected by difference between the
incomplete blocks.
proc glm;
class rep bl tre;
model res=rep bl(rep) tre ;
run;
The ANOVA result for same data set using SAS is:
69
Error 16 1.23680741
In the recent version of SAS a procedure for lattice has been developed and the following
command may be used as an alternative.
Note that the way the program is written has an assumption of how the data should be
entered, since we do not specify the block and rep factors in the proc statement. The data
should be read as follows. Only the input statement is given here and the reader can read
the data from a file or can use cards to enter directly as usual.
The following proc glm program gives same result as that of proc lattice:
Proc GLM
Class rep bl tre;
Model res = rep tre bl (rep);
The difference of this GLM program from the previous one is the position of the “tre”
treatment. Here “tre” is put after “rep” so it is not adjusted.
The reason for using two procedures for same analysis is to show users the power of
GLM procedure. Thus with GLM procedure we could set a complete analysis for lattice.
The efficiency could be calculated as indicated in step 6 of previous page,
70
Comparison of some important statistical packages using their output of the Lattice
design is necessary at this point as there are some practical limitations in the use of out
puts from such packages.
Example data was taken from Ato Gelana Soboka’s MSC thesis work entitled “Heterosis
and combining ability between CRMMYT’S and Ethiopian MG”. We are grateful to him
for allowing us use this data set as an example. The experiment was designed as simple
Lattice with 2 replications and a square incomplete block of 10 x10 (100 entries). The
experiment was done at 3 locations.
The data was analyzed using MSTATC, SAS and Agrobase and the results compared.
Result of Mstatc
MSTATC produces treatment (adjusted and unadjusted SS), block within reps (adjusted
for treatment), three types of error, efficiency, LSD, CV and adjusted entry means. The F
– test for entries is the ratio of treat SS to effective error SS. In the example below
efficiency of Lattice was found to be about 110.44 %
71
Data file : ALBK1_
Title : Heterosis and combining ability between CIMMYT's and
Ethiopian MG
Function : ANOVALAT
Data case no. 1 to 200
Variable number 4
yield kg/ha (12.5%m)
A N A L Y S I S O F V A R I A N C E T A B L E
Variable number 4
yield kg/ha (12.5%m)
AGROBASE
Agrobase also gives treatment (adjusted and unadjusted), adjusted block (within rep) and
intra block error effects. It also gives CV, R2 , RE , adjusted entry means and different
standard errors. The good things about Agrobase is that it calculates the adjustment
72
factor µ which we saw in previous hand calculation example, so that we can determine
the level of balance of the design. Unlike MSTATC, Agrobase automatically determine
the arrangement and report the repetitions rather than requesting from user.
There is a wide difference in the output between the two packages. To start with, adjusted
treatment SS for MSTATC is 191328871.03 (with 99 df ), while this value is
159135138.5 ( with 99 df) for Agrobase. The difference between adjusted and unadjusted
treatment SS for MSTATC is so small, while for AGROBASE it is large. This is
possibility because the two packages based their principle of adjustment on different
references. For example, MSTATC seems to have used a rough approximation of
treatment sum of square from adjusted treatment totals which has been adopted by
cochran and cox in 1957. Cochran and cox argued that since there was not perfect
adjustment factor and since unadjusted treatment means contains block effect, comparing
it with intrablock error is not correct. Thus they roughly adjusted treatment SS from
adjusted totals and used effective error as a divisor for F – tests. On the other hand
AGROBASE is based on recent development in the incomplete block and used the extra
sum of square principle (Mead,1988) to adjust treatment sum of squares. Thus, no
effective error SS was calculated here and all the test was done on the intrablock SS. The
relative efficiency of Lattice is also calculated as a ratio of RCBD error to intrablock error
SS. For this and other reasons the adjusted entry means area also different for the two
packages.
ANALYSIS OF VARIANCE
SAS
SAS is a general purpose program and the output expected from SAS depends on how
you supplied the program to the package and which sum of square you use since it do not
strictly follow a pre-determined format of output for lattice as in MSTATC or Agrobase,
when proc GLM is used. When proc lattice is used, however, the output resembles that
from the two packages we have seen. In incomplete block design, the order of entering
73
the factors in the model specification in SAS is most important when proc GLM is used.
For example the two orders: 1) Rep, treat, block (rep) and 2) Rep, block (rep), treat do not
produce same result for type I sum of square, because for (1) treatment is not adjusted
but for (2) treatment is adjusted for block (within rep). But if we consider results of type
III sum of square both (1) and (2) are the same because in type III SS each effect is
adjusted for all other effects included in the model. For example when (2) was fitted,
block (rep) SS (type I) was 73120835.5, but for type III SS, the result is 30837048.1
because type III SS adjusts the effect for all other factors included in the model
irrespective of the order of entry to the model.
Regarding the analysis result, SAS and AGROBASE agrees very much, as adjusted
treatment mean and block (rep) is very close for the two. The error sum of square that
SAS supplies is the intrablock error and the tests could be done using it.
The next question is how do we analyze lattice for more than one environment (Year or
location). Obviously both MSTATC and Agrobase can’t handle this. But SAS can easily
handle through similar steps. All the 3 location of this data set was combined using the
following procedure in old SAS version and results obtained.
Proc GLM;
Class rep block Loc treat;
Model yield = Loc rep(Loc) bloc (rep) treat treat*Loc;
Run;
ANOVA
Source Df SS MS F Prob
Locati 1 108620.7 108620.7 0.011 <.0001
Rep(Loc) 2 75608772.3 37804386 37.03 <.0001
Bloc(rep) 18 30006166.3 1667009.2 1.62 <.0561
Treat 99 316291936.8 3194868.0 3.13 <.0001
Treat*Loc 99 174673488.7 1764378.7 1.73 <.0008
Intrablock 180 183768312 1020935.1
error
Note: Here rep is assumed to be nested under location and Block is also assumed to be
nested under replication.
74
5.3. Alpha Designs (Generalized Lattice)
Generalized lattice, or alpha designs (Peterson and Williams, 1997) provide lattice (with
some loss of balance and efficiency) where conventional square or rectangular lattices can
not be generated, such as 29 treatments in two complete blocks.
α(0,1), α(0,1,2). . . . .
The notation α(0,1), α(0,1,2) etc… refers to occurrence of treatments in a block. For
example α(0,1) is a symbol which says that some treatments occurs only once in a block
with others, while others do not occur together with other in a block at all.
The development of alpha designs removed the restrictions on the number of treatments
to be considered and its relation with block size required for lattice designs. To
demonstrate randomization of the design by hand, consider the comparison of 28 (p = 28)
treatments where the block size is 4(r = 4), i.e., p is multiple of r. Here p = sr, and we
write down the four sets of s = 7 treatments
Then as for a lattice design, we construct blocks in successive replicates so that (i) each
block includes one treatment from each set; and (ii) for each replicate the treatments in
each block have not occurred together in a block in any previous replicate.
For the first replicate, we choose the 7 blocks to be, I: the first treatment of each set; II:
the second treatment of each set; VII: the seventh treatment of each set. For the second
and subsequent replicates we rotate the list of treatments in the several sets so that the
sets occur in different positions relative to each other. For example, the second
replication might be based on the rearrangement:
75
Set 1 Set 2 Set 3 Set 4
1 9 17 26
2 10 18 27
3 11 19 28
4 12 20 22
5 13 21 23
6 14 15 24
7 8 16 25
Set 1 is unchanged, set 2 is moved up one position, set 3 is moved up two positions and
set 4 is moved up four positions. The blocks for the second replicate are formed from
the first, second,..., seventh positions in the cycled sets.
For the third replicate, set 2 is moved up three places from the original position, set 3 six
places and set 4 five places:
The number of places by which each set is rotated are chosen so that each pair of sets
occurs in different relative positions in each replicate (e.g. for sets 3 and 4 , in the first
replicate treatment 15 occurs with treatment 22, in the second replicate with treatment
24, and in the third with treatment 28). The resulting design is shown in table 5.3.
76
Table 5.3 Alpha design for comparing 28 treatments in blocks of 4 plots
Replicate 1
Block I 1 8 15 22
Block II 2 9 16 23
Block III 3 10 17 24
Block IV 4 11 18 25
Block V 5 12 19 26
Block VI 6 13 20 27
Block VII 7 14 21 28
Replicate 2
Block VIII 1 9 17 26
Block IX 2 10 18 27
Block X 3 11 19 28
Block XI 4 12 20 22
Block XII 5 13 21 23
Block XIII 6 14 15 24
Block XIV 7 8 16 25
Replicate 3
Block XV 1 11 21 27
Block XVI 2 12 15 28
Block XVII 3 13 16 22
Block XVIII 4 14 17 23
Block XIX 5 8 18 24
Block XXI 6 9 19 25
Block XXI 7 10 20 26
77
Table 5.3.1. Alpha design for comparing 50 treatments in 3 replicates, each replicate
consisting of 10 blocks of 5 plots.
1 45 22 21 5 25 31
2 34 39 2 11 48 50
3 8 26 19 41 35 3
4 17 31 16 40 9 42
5 42 4 31 3 32 14
6 5 36 7 35 26 17
7 28 24 46 17 49 12
8 25 44 39 38 41 16
9 14 38 25 13 23 19
10 29 3 45 23 33 46
11 12 27 20 12 13 19
12 43 50 32 26 6 18
13 33 23 18 6 7 8
14 48 15 50 14 15 36
15 21 46 34 1 37 29
16 10 7 4 36 22 2
17 32 30 8 47 28 47
18 37 41 44 15 40 43
19 11 9 33 48 34 45
20 47 16 37 22 21 24
21 49 2 49 42 5 39
22 1 18 28 43 38 27
23 19 6 10 27 30 44
24 20 40 24 29 4 11
25 13 35 9 30 20 1
78
Alpha designs are most effective when the block size, r, is less than the square root of the
number of treatments, p, and hence less than s. When r > s, some treatment pairs occur
together in a block in more than one replicate, which can lead to unequal levels of
precession for some comparisons (i.e., some treatment comparisons are made with greater
precision than other treatment comparisons).
Particular set of rotations for the sets for each replicate is important. Some rotation
patterns provide more efficient designs than others. A computer program, ALPHA
provides optimal designs for trials with up 500 treatments (Journal of Agricultural
Science (1978), Vol. 90, pp.395-400, provides optimal designs for trials with up to 100
treatments.)
REML provides a fully efficient analysis, but two aspects of fitting models are not
explicit in the results. First we should check that the degrees of freedom on which
variance components are estimated are sufficient for reliable estimates; minimum of 8 d.f.
is a safe limit. Second, the process of selecting a model itself introduces a small
additional imprecision into the estimation procedure, which is not reflected in the
standard error of treatment differences from the REM analysis.
Example for Alpha design is given on Table 5.3 . The data is the yield obtained by
applying 28 treatments.
Here we applied the GLM procedure to roughly estimate the ANOVA so that every body
can use since the earlier version of SAS do not support REML procedure.
REPLICATION I
Block 1 30 19 26 18
Block 2 28 26 17 29
Block 3 31 27 30 20
Block 4 15 27 31 26
Block 5 22 16 19 27
Block 6 32 22 19 21
Block 7 22 30 31 20
79
REPLICATION II
Block 8 31 31 25 30
Block 9 33 22 26 19
Block 10 19 16 28 34
Block 11 23 24 27 14
Block 12 27 24 20 34
Block 13 28 28 19 30
Block 14 26 27 17 19
REPLICATION III
Block 15 31 20 16 21
Block 16 25 20 35 26
Block 17 21 29 25 30
Block 18 24 21 34 25
Block 19 18 25 28 32
Block 20 24 34 25 23
Block 21 25 29 31 37
The SAS program applied for the above data is (by omitting the data step):
proc glm ;
class blo repl trea;
model yi=repl blo(repl) trea;
Run;
Dependent Variable: YI
As in Lattice design, here block with rep and treatment effects are both adjusted for each
other, and the error term is intrablock error. SAS analyzes most incomplete design with
the GLM procedure as long as the incompleteness is so serious that terms could not be
easily estimated. The REMEL procedure refine the analysis through the variance
component technique. For this example, none of the terms are significant at 5% level.
80
5.4. Cyclic Designs
They are family of incomplete block designs which are based in cyclic development of
one or more initial blocks. They are basically developed for ease of allocation of
treatments to blocks, in case of incomplete blocks- where keeping balance is often a
difficult exercise. All BIB and PBIB designs obtained through the method of differences
are called cyclic designs. If there are v treatments, block of size k and treatments
replicated r times, then the C(v,k,r) can be generated.
Example
Let treatments, V = 6, block size k = 6 and treatments applied are replicated r = 6 times.
❶ If the initial block is (0,1,2), then successive blocks are : (1,2,3), (2,3,4),
(3,4,5),(4,5,0),(5,0,1) α (0,1,2)
Extraneous effects, such as harvesting, lodging, slope and wind can be confounded with
rows and affect yield estimate
Example
Consider seven treatments in seven blocks of three units each. (the case of one way
blocking);
81
❶ Incomplete Block (IB) Design
Block
1 2 3 4 5 6 7
A A A B B C C
B D F D E D E
C E G F G G F
This is balanced design because each treatment occurred with each other in a block. For
example, if you observe treatment A, we see that A occurred with B and C in Block1,
with D and E in block 2 and with F and G in block3.
Column
1 2 3 4 5 6 7
1 A D F B G C E
Rows 2 B A G F E D C
3 C E A D B G F
Although the design was balanced when subjected to one way blocking it is now
partially non-orthogonal and unbalanced when row and columns are considered.
The SAS program computing the hypothetical data is:( the data is shown in the ‘cards’
statement)
82
The SAS output is:
Such kind of designs can’t be analyzed with MSTATC or even in Agrobase. MSTATC
can only manage Lattice design among the incomplete blocks. The SAS result shown
above is adjusted treatment and column effects. In SAS, the three types of sum of square
are to simplify some of the difficulties related to adjustments. Type one sum of square in
SAS produce SS in the order of fitting, that is any effect following another effect is
adjusted for the preceding effect. But type III SS adjusts each effect for all other effects
irrespective of the order of fitting. Hence, when there is a considerable non-
orthogonality, SS obtained when effects are fitted in different order is different.
83
5.6 Unreplicated Design
Objectives in experimental design are:
Thus, several materials may enter into the initial stages of breeding line and there is no
sufficient resource to sufficiently replicate all entries for statistical analysis. On the
other hand, it is not possible to get a land which is uniform enough to enable us assume
that all entries are applied on homogenous plots and can be compared directly.
Therefore, it is important to look for ways of either control soil heterogeniety or develop
appropriate design.
There are several ways of overcoming the problem, the most useful ones are:
By using only test genotypes in the field, we compare only genotypes within
hypothetical square blocks. It looks fair enough. The problem, however, is that it is
not possible to compare all of them at once and you may have to select several
entries.
One replicate of a design is used in each site. This is even reasonable, but the
problem often is that genotype by environment interaction could be a limiting factor
for successful comparison.
Since the difficulty is to estimate error mean square from replicated plots to be used in
F-ratio, the other and most commonly used alternatives among breeders is replicate
checks only in each block and calculate error mean square from them. This will lead us
to the family of Augmented designs.
84
5.6.1 Augmented designs, developed by Federer (1956)
The principle followed in this design is as follows:
Example
Block
1 2 3 4 5 6
A C 21 A B A
13 17 A 2 C 19
8 9 C B 12 C
B A 15 10 16 20
C 11 B C 5 B
18 8 23 3 6 4
7 24 1 22 A 14
B1 B2 B3 B4 B5 B6
Since all new selections did not occur together in a block, a direct comparison of
them is not possible, but since checks occurred together in a block it is possible to get
a reasonable estimate of block effect so that it will be used in adjusting treatments.
The adjustment index is calculated as follows:
85
ANVOVA for checks
Source df
Block b-1
Checks c-1
Residuals (b-1)(c-1)
It should be noted that the precision of residual mean square depends on the number of
check plots you included per block. Include as much checks as possible given the
resource and the material. Particularly, checks below 3 are not acceptable in trials with
augmented designs.
Example
Analysis of the standard Augmented Design of the above example using Agument
software.
86
The four standard errors are:
Variable: HD
ANALYSIS OF VARIANCE:
The above ANOVA can also be constructed using hand calculation by only considering
the checks as treatments and blocks. That means since we have four check varieties the
data can be analyzed based on RCBD with four treatments.
For MAD-I the sub-plots should be square, or almost square, so that sub-plots are equi-
distant from the central plot. The 3x3 squares are planted such that the central plots are
laid out as a Latin square, with orthogonal rows and columns, to give a bi-directional
estimate of soil heterogeneity. To estimate sub-plot error, two whole plots are randomly
chosen for each control variety and each control variety is thus randomly assigned.
Modified augmented design type 2 is more commonly used than MAD-1, because MAD-
2 is meant for field layout with long, rectangular sub-plots, typical of small grains
breeding programs.
87
Example
∗ How to distribute the check genotypes in the field?
∗ What should be the frequency of the checks?
GOAL: To obtain the most accurate ESTIMATE of row and column effects in order to
make the most accurate estimate of entry mean yields.
Figure 5.6.2: Design plan for MAD-I with 40 check plots and 120 entry plots.
2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
A B A B
A B A B
A B A B
A B A B
B A B A
B A B A
B A B A
B A B A
A A B
B B
Source Df
Total 40
Correction for MEAN 1
Row (ignoring checks and columns) 9
Column (ignore checks, Eliminating column) 15
Check , (Eliminating row and column) 1
Error 14
The modified augmented design is well furnished in Agrobase package. However, it can
be analyzed by any other packages.
To illustrate the use of modified augmented design II (MAD-II)data was taken from
AGROBASE training material and analyzed as follows. The data set is also attached to
help readers understand the design and apply to their own field of research.
Note how the data is arranged. The central plot is always occupied by the control plot.
Cp = Control plot
Csp = Control sub-plot
88
Fig. 5.6.2.1. Plan for MAD-II with 16 control plots and control sub-plot, and 48 test
entries.
Column
1 2 3 4
1 3 5 B
Row 1 2 A 6 9
C C C C
B B 7 A
A 4 8 10
23 19 15 11
2 24 20 16 12
C C C C
A 21 17 39
B 22 18 40
25 29 33 37
3 26 30 34 38
C C C C
27 31 35 39
28 32 36 40
A 45 A B
4 47 46 43 A
C C C C
48 B 44 41
B A B 42
Note that intersection of row by column is known as whole plot. Therefore, there are 16
whole plots, each whole plot is split in to 5 sub-plots. The central plot is always
assigned to control plot. Here since the sub- control plots are not assigned to control
varieties at each whole plot, the test entry occupies 48 plots of the 80 plots available.
Had there been two control sub-plots per whole plot, we could use only 32 entries for
this experiment.
89
Table 5.6.2 : Data file for modified augmented design.
Entry Name Plot Row Col Cp Csp Yld
1 ENTRY -1 1 1 1 56.91
2 ENTRY -2 2 1 1 57.81
9010 CONTROL 3 1 1 1 58.25
9033 SUB-CONTROL B 4 1 1 2 57.7
9023 SUB-CONTROL A 5 1 1 1 55.84
3 ENTRY - 3 6 1 2 61.08
9022 SUB - CONTROL A 7 1 2 1 59.84
9010 CONTROL 8 1 2 1 61.78
9032 SUB- CONTROL B 9 1 2 2 60.46
4 ENTRY - 4 10 1 2 61.16
5 ENTRY - 5 11 1 3 66.69
6 ENTRY - 6 12 1 3 64.93
9010 CONTROL 13 1 3 1 65.57
7 ENTRY - 7 14 1 3 63.64
8 ENTRY - 8 15 1 3 61.93
9038 SUB - CONTROL B 16 1 4 2 69.15
9 ENTRY - 9 17 1 4 67.75
9010 CONTROL 18 1 4 1 67.7
9028 SUB-CONTROL A 19 1 4 1 69.2
10 ENTRY - 10 20 1 4 69.45
11 ENTRY - 11 21 2 4 72.76
12 ENTRY - 12 22 2 4 72.48
9010 CONTROL 23 2 4 1 72.44
13 ENTRY - 13 24 2 4 70.15
14 ENTRY - 14 25 2 4 72.50
15 ENTRY - 15 26 2 3 65.46
16 ENTRY - 16 27 2 3 64.87
9010 CONTROL 28 2 3 1 67.6
17 ENTRY - 17 29 2 3 67.73
18 ENTRY - 18 30 2 3 71.37
19 ENTRY - 19 31 2 2 63.86
20 ENTRY - 20 32 2 2 68.48
9010 CONTROL 33 2 2 1 66.03
21 ENTRY - 21 34 2 2 62.23
22 ENTRY - 22 35 2 2 67.36
23 ENTRY - 23 36 2 1 59.57
24 ENTRY - 24 37 2 1 57.08
9010 CONTROL 38 2 1 1 58.46
9027 SUB - CONTROL A 39 2 1 1 61.56
9037 SUB - CONTROL B 40 2 1 2 58.89
25 ENTRY - 25 41 3 1 65.29
26 ENTRY - 26 42 3 1 64.24
9010 CONTROL 43 3 1 1 60.5
27 ENTRY - 27 44 3 1 61.21
28 ENTRY - 28 45 3 1 61.78
29 ENTRY - 29 46 3 2 64.29
30 ENTRY - 30 47 3 2 66.35
9010 CONTROL 48 3 2 1 67.61
31 ENTRY - 31 49 3 2 68.33
32 ENTRY - 32 50 3 2 67.27
90
33 ENTRY - 33 51 3 3 67.46
34 ENTRY - 34 52 3 3 73.28
9010 CONTROL 53 3 3 1 72.77
35 ENTRY - 35 54 3 3 70.69
36 ENTRY - 36 55 3 3 70.64
37 ENTRY - 37 56 3 4 77.13
38 ENTRY - 38 57 3 4 79.02
9010 CONTROL 58 3 4 1 73.38
39 ENTRY - 39 59 3 4 76.2
40 ENTRY - 40 60 3 4 73.01
9031 SUB-CONTROL B 61 4 4 2 79.07
9021 SUB - CONTROL A 62 4 4 1 79.01
9010 CONTROL 63 4 4 1 78.2
41 ENTRY - 41 64 4 4 76.02
42 ENTRY - 42 65 4 4 79.2
9025 SUB - CONTROL A 66 4 3 1 75.18
43 ENTRY - 43 67 4 3 72
9010 CONTROL 68 4 3 1 72.33
44 ENTRY - 44 69 4 3 76.14
9035 SUB - CONTROL B 70 4 3 2 71.32
45 ENTRY - 45 71 4 2 69.04
46 ENTRY - 46 72 4 2 73.07
9010 CONTROL 73 4 2 1 67.7
9036 SUB - CONTROL B 74 4 2 2 73.31
9026 SUB - CONTROL A 75 4 2 1 72.59
9024 SUB - CONTROL A 76 4 1 1 64.61
47 ENTRY - 47 77 4 1 67.59
9010 CONTROL 78 4 1 1 69.75
48 ENTRY - 48 79 4 1 68.02
9034 SUB - CONTROL B 80 4 1 2 64.34
There are 16 control plots in the whole experiment and the preceding ANOVA is based
on those plots, therefore whole plot errors are estimated from control plots. Sub-plot
error is estimated from the residual of RCBD ANOVA on whole plots containing sub-
plot controls considered as “blocks” and the control varieties as “ treatments”.
Adjustment for difference in row and column is also required. There are two types of
yield adjustment, but we will discuss what is called method 3 adjustment. For this
method the adjustment factor is regression coefficient, b. Once this value is determined,
yield can be adjusted as follows:
Where Yijk = value of control plot in ith row and jth column line in the ijth whole plot
C = control line
b = regression coefficient
In the ANOVA below, unadjusted mean of control plot is given to be 67.504 and
regression coefficient of 1.02.
91
Output from the Agrobase package is given as follows.
ANALYSIS OF VARIANCE
Modified augmented design (2) : Control Plots ANOVA
File: C:\AGRO97\DEMO\DEMAUG22.DBF
No. 1: 67.504
Same analysis could be obtained using other packages. But, more steps may be required
to rearrange the data set to suit the software to be used. The first activity is to collect the
control plot yield in a 4 x 4 row and column (table 5.6.2.2). Then, it turn out to be just a
row and column design without a treatment factor. Yield will be the response variable.
Table 5.6.2.2: Control plot yield
Column
1 2 3 4
Row 1 58.25 61.78 65.57 67.7
2 58.49 66.03 67.6 72.44
3 60.50 67.61 72.77 73.38
4 69.75 67.70 72.33 78.20
To enter the data in to computer create 3 variables, ROW, COLUMN and YIELD as
follows:
92
To analyze using MSTATC, select FACTOR menu, then select “CRD Two Factor” sub-
menu. To analyze same data in SAS use the following program:
PROC ANOVA;
Class ROW COLUMN
Model YIELD = ROW COLUMN;
Run;
Both MSTATC and the SAS program (above) give the same result as the one obtained
from Agrobase.
93
6. Treatment structure
6.1 Factorial experiments
So far we covered four families of experimental designs. These designs are applicable to
any type of experiment, regardless of the structure of treatments. Now we will look at
two types of experiments. Single factor and multi - factor experiments.
Examples we have been using so far concerns a single factor experiment. Single Factor
experiment means that the investigator is concerned in testing several levels of one factor
while keeping all other factors at a constant level.
For example, in variety trail we cited earlier, the aim was to compare the yielding ability
of different varieties under uniform management and cultural practices. Variety is
considered as a one factor, while the different varieties are considered as different levels
of the factor.
To widen the scope of an experiment, it is a common practice, these days, to include two
or more factors so that their effect and cross-effects are tested simultaneously: for
example testing five varieties under 3 nitrogen levels. This is commonly known as
factorial experiment (Which we will see later on).
The experimental designs discussed so far are useful to study the effect of a single factor
on the response variable. In some experiments it becomes necessary to study the effects
of two or more factors. For example, a plant breeder might be interested in measuring
the effect of variety and planting date on the yield of wheat. Such experiments that
consist of 2 or more combinations of different factors are referred to as factorial
experiments. In factorial experiments the treatments are compared on the basis of the
effects of the individual factors and their interactions.
There are advantages and disadvantages in using factorial experiments. If the purpose of
the experiment is to investigate the effect of each factor, rather than the combination of
levels of the factors that produce a maximum response, then separate experiments could
be conducted where each deals with a single factor.
Usually factorial experiments are more efficient than single-factor experiments but they
can lead to an increase in complexity, size and cost of an experiment especially if some
of the factors included are less important. Therefore, it is preferable to use an initial
single factor experiment, especially if the number of potentially important factors is
large.
Another disadvantage of factorial experiments lies in the less precise estimates that result
from problems associated with heterogeneity of factors as the number of treatment
increases.
The advantage of factorial structure, over single factor experiment, is that it will enable
us to simultaneously test effects of several factors and their cross effects known as
94
interactions. Interaction provides a measure of the pattern of the effect of one factor over
the other factor.
Consider two varieties (V1 and V2) and two Nitrogen levels (N1 and N2). Consider mean
yield of a1 for V2 when N1 is applied, and b1 when N2 is applied, and mean yield of a2
for V1 when N1 is applied and b2 where N2 is applied.
The graph below shows that Nitrogen increase the yield of both varieties, V1 and V2 at
different rate. This means that there is an interaction between variety and nitrogen level.
Consequently, the two lines seem to cross each other at some point if they are extended.
For this example, the interaction is described by the difference between (a1 –a2) and (b1-
b2). Obviously, as we can see from the graph the distance between b1 and b2 is much
smaller than the distance between a1 and a2, hence there is interaction between N and V.
Thus, we say that nitrogen increases yield of both varieties but application of N2 boost
yield of V1 whereas, N2 increases yield of V2 slightly over N1, since the difference in
yield between the two varieties is small for N2
b1
a1 * V2
* b2 * v1
a2
*
N1 N2
In the following graph, the two lines are parallel measuring that nitrogen increases yield of both varieties at
the same rate. Which implies no interaction between them.
Here, unlike the above graph, the distance between a1 and a2 is nearly the same as that between b1 and b2.
Thus, we may say that the difference between the two varieties remain the same irrespective of the type of
fertilizer applied. Hence, variety do not depend on the type of fertilizer applied.
b1 V 2
*
a1 b2 V1
* *
a2
*
N1 N2
The following table shows the treatment structure of a 2-factor experiment, the effect of
fiber (factor A) and protein (factor B) in a diet. The levels of fiber are three, f0, f1, and f2
(i.e., low fiber, medium and high). On the other hand, protein has two levels, p0 and p1
(low and high). This can be represented by a two-way table where the cross classification
of each level of one factor occure with each level of the other.
95
Factor B (protein)
P0 P1
f0 F0p0 f 0 p1
Factor A
(Fiber) f1 F1P0 f 1 p1
f2 F2p0 f 2 p1
Factorial experiments may be laid out in any convenient design such as CRD, RCBD,
Latin square, etc.
For example, if CRD is to be applied in the above study, the individual treatment
combinations, f0p0, f1p0, f0p1, f1p1, f2p1 and f2p1, are regarded as treatments and
experimental units (animals ) are randomly assigned to each of them.
The resulting ANOVA will provide tests for the significance of the difference among the
proteins, the difference among the fibers, or the significance of the protein X fiber cross
effect (interaction).
Example : The nutritive value of a certain edible fruit was measured in a total of 24
specimens representing six specimen of two varieties grown in each of two geographic
regions taken at random . The results were as follows.
II 8.9 9.1
9.2 13.1
5.2 13.2
7.7 8.6
7.8 9.8
5.7 9.9
Test for a significant difference between regions, varieties and their interaction.
Here no blocking is used, hence the design adopted is completely randomized design
(CRD). Calculation of Analysis of variance can be done in two ways I) By hand or II)
using computer programs as follows:
96
Manual calculation.
Note that the multipliers given are coefficients for main factor effects. For example the
coefficients for A (-1 +1-1+1 ) finds the difference between the two levels of A for each
level of variety. The factorial effects is thus calculated easily by multiplying treatment
totals by their respective coefficients.
ANOVA table
Presentation of results:
There is an evidence of differences between the regions and between the varieties. The
evidence for differences between varieties is very strong compared to that of region.
Here since both factors have only two levels it is already clear from this test which levels
are higher than others and no furthers contrasts are important. On the other hand, the
interaction effect is not significant at 5% level. Hence, varieties are not associated with
regions. That means regions treat both varieties equally. Therefore, presentation of the
main effect means alone is important, since no cross effect exists.
97
Table of Means
Levels
1 2
Region 10.79 9.02
Factory
Variety 7.83 11.79
From the table above, region 1 is better yielding than region 2 and variety 2 performs
better than variety 1.
Same analysis could be done with computer. For example, MSTATC does this analysis
using “CRD with 2 factor” sub-menu under the “Factor “ menu. SAS has been used to
do analysis of same data set as follows:
Proc anova;
Class geo var yield;
Model geo var geo*var;
Run;
Let’s us now consider another more complicated example, and discuss presentation
styles in detail:
98
analysis. Note that most statistical packages could be used to do analysis of this
experiment (for example, in MSTATC, “RCBD three factor” sub-menu under “Factor”
menu could be used.)
ANOVA Table
Source df MS
Block 2 234
Fertilizer (F) 2 105.3
Variety (V) 3 79.5*
Managnt (M) 1 128.5**
FxV 6 466.0***
FxM 2 93.6*
VxM 3 28.3
FxVxM 6 16.4
Error 46 26.2
Total 71
Now the major concern here is which table of mean to present because we have both
main effects and interaction means. As clearly depicted in ANOVA table above all main
effects and two interactions are significant at 5% level. If we consider F, V and F x V all
of them are significant, but since differences between response of say V is highly
affected by the level of fertilizer applied, the main effect means do not give much
information. That means we can’t conclude, for example, by saying that variety A is
different from others, because since interaction effect is significant there is an indication
that variety A might be different from other varieties for some fertilizer levels and not
for others. Therefore, we need to present interaction mean only (F x V). Regarding F
and M, however, the main effects are significant but interaction is not therefore it is
better to present the main effect means alone. In situations where main effects are highly
significant (say at 0.1% level) and interactions are significant at 5% level, it is good idea
to present both main effect and interaction means.
These designs are also called nested designs, because of repeated sampling and sub-
sampling that occur in the design. Since more independent measurements are taken for
the subplot factor than the main plot factor, the subplot factor is measured more precisely
than the main plot factor, hence comparison between sub-plot factor are more precise.
Example
Suppose that we study the properties of 3 varieties of cotton breed for resistance to wilt
99
as the subplot factor and 4 dates of sowing as the main plot factor, and the experiment is
conducted in 4 replications.
1. Divide the experimental area into 4 blocks( r =4). Further divide each block into 4
main plots as shown in Figure 6.1.
1
Main plot 1
Main plot 1
Main plot 1
Main plot
Figure 6.1 Layout of an experimental design into 4 blocks and 4 main plots
2. Randomly assign the main plot treatments Factor A, (4 dates of sowing: D1, D2, D3,
and D4,) to the main plots among the 4 blocks or replications as shown in Figure 6.2
D4 D2 D1 D3 D1 D4 D2 D3 D3 D2 D4 D1 D1 D4 D3 D2
Figure 6.2 Random assignment of 4 dates of sowing the four main plot in each of
the 4 replications
3. Divide each of the main plots in to 3 subplots and randomly assign the cotton
varieties (Factor B) to each subplot as shown in Figure 6.3.
100
each replication.
We can see that the number of times the subplot factor is tested is much larger than that
of the main plot factor. Therefore, the subplot factor is estimated more precisely than the
main plot factor, and care should be taken when deciding on the factor to be assigned to
the subplot. Also note that the major difference between RCBD and split plot lies on
the concept of randomization. The dates of sowing are randomized on the main plots
only, while the varieties are randomized on the sub-plots within main plot. In the case of
RCBD, it is a factorial combination of date by variety which is randomized within a
block. For example, since we have 16-treatment combination of date by variety, we
normally randomize them on to 16 plot within each block. In the case of split plot since
we have two sized plots and randomization is being done at two stage, we have two
different error terms. One of the problems in these areas that some researchers may
initially design their experiment as split plot but analyze as RCBD. This automatically
ignores the difference in randomization between the two and the split plot treatment will
be affected while the main plot treatment is falsely boosted.
Analysis
Model : Xijk = µ + αi + βj(i) + ε(ij)k
Using the least square estimators for parameters, transposing 8 … to the left,
squaring and taking the sum over all the subscripts we get.
SSA = ∑8i2… - CF
bn
SSB = ∑8j2.. - CF
an
SSBA = ∑∑Xij. 2 – CF
n
101
The ANOVA table
Source d.f SS MS F
Replication r-1 rss
A a-1 SSA SSA/A-1
Error (a) R(a-1) Err SS ErrSS/r(a-1)
B b-1 SSB SSB/ b-1
BxA (a-1)(b-1) SSBA SSA x B/(a-1)(b-1)
Error (b) ab(n-1) SSE SSE/ab(n-1)
Total abn-1 TSS
Many statisticians and senior researchers, however discourage the use of split plot for a
number of reasons:
1. In split plot the sub-plot factor is more precise in the expense of main-plot factor and
this will have a due influence on our conclusion regarding the main plot factor.
2. Split plot have about 5 s.e’s for comparing the levels of different factors, and people
often make a mistake when supplying error mean square for performing multiple
range tests in MATATC (see later).
3. The main plot error often associated with small degree of freedom which may be
open to criticism. But, to increase the df one may have to select several levels of a
factor which is both non-economical and not practical. Therefore, there is no need of
complicating things, unless practical applications of treatments require such
differential plot size, when the use of RCBD will equally do an excellent job.
Practical necessity occures in situation like irrigation trial or method of ploughing
where large plots are practically inevitable as the treatments can not be done on small
plots.
Example
102
N rate Kg/ hectare
Replication Hybrid 0 35 70 105
Yield bu/ hectare
I H1 130 150 170 165
H2 125 150 160 165
H3 110 140 155 150
H4 115 140 160 140
H5 115 170 160 170
Source d.f. SS MS F
Total 39 20,244.38
Replication 1 4,100.63 4,100.63
Nitrogen(A) 3 12,015.88 4,017.29 42.76**
Error (a) 3 281.87 93.96
Hybrid(B) 4 2,466.25 616.56 20.56
Nitrogen X Hybrid
(A X B) 12 863.75 71.98 2.40ns
Error (b) 16 479.97 29.99
The interactions between Nitrogen rate and hybrid were not significant. The differences
between the N rates are significant at 1% level. This shows that there is strong evidence
that grain yield was significantly affected by nitrogen. Similarly, the results indicate that
in both replications grain yield was significantly affected by the hybrid.
103
Proc GLM;
Class Replicat Nitrogen Hybrid yield;
Model Yield = Replicat Nitrogen Replicat *Nitrogen Hybrid Hybrid* Nitrogen;
Test H = Nitrogen e = replicate * Nitrogen;
Test H = replicate e = Replicate * Nitrogen;
Means Nitrogen Hybrid;
Run;
Level of ------------YIELD------------
NITROGEN N Mean SD
0 10 129.000000 13.0809446
35 10 156.500000 13.3437459
70 10 171.000000 13.7032032
105 10 172.000000 19.3218357
Level of ------------YIELD------------
HYBRID N Mean SD
1 8 161.875000 21.8660696
2 8 161.250000 22.1601315
3 8 148.125000 19.9888362
4 8 147.500000 20.3540098
5 8 166.875000 27.5081157
One of the disadvantages of using split-plot design is that there are several s.e to use for
comparing the levels of the different factors. One mistake that people often do in
applying LSD or Duncans multiple range test in split-plot is that they supply error (b) to
MSTATC so that the program compute and test mean differences for main plot, split plot
and as well as their interactions. As you might see below s.e for comparing two levels of
main plot factor for same or different level of sub-plot factor is even more complicated.
Therefore, to do multiple range test in MSTATC supply the relevant error mean square.
To compare levels of main plot factor, for example supply error (b) with n = rxa. To
104
compare levels of split-plot factor at one level of main plot factor supply errors(b) with
n = r etc..
When more than two factors are available, there are different possibilities for designing
split-plot. When 3 factors are included, two factors may be applied on main plot and the
third factor on split-plot or visa vis. Consider the case of 3 factor, where two levels of
fertilizer, two management techniques and two varieties were compared in split-plot
design with 3 replications. It was desired to put both fertilizer and management
technique on main plot and variety on sub-plot. Experimental plan for the first rep is
given below. Four main plots are required because we have 2 x 2 = 4 Fertilizer-
management combination and under each main plot we require 2 sub-plot to match with
the levels of variety.
Fig 6.1.1: Experimental plan for 3-factor split-plot (for first rep) with fertilizer and
management factors on main plot and variety on sub-plot.
An artificial data was created and analysis done using this design, and the result from
MSTATC output is given below.
A N A L Y S I S O F V A R I A N C E T A B L E
105
Coefficient of Variation: 31.34%
Note the analysis of variance structure. Fertilizer, management and their interaction are
all in the main plot stratum, while all others are in the slit-plot stratum. When standard
error are calculated, different error mean square are used. For example, s for mean
y
group 2, group 4, and group 6 used error (a), while s for group means 8,10,12 and 14
y
are used error (b). If, for example, comparison of levels of variety at a fixed level of
fertilizer and management is required then both error (a) and error (b) will be used as
shown in calculation of different s.e’s in previous page.
In measuring the interaction effect between 2 factors, the experimental area is divided
into 3 plot. We allocate factor A and B, respectively, to the vertical and horizontal-strip
plots, and allow the interaction plot to accommodate the interaction between these 2
factors. The vertical and the horizontal plots-are perpendicular to each other. However,
in the strip-plot design the relationship between the vertical and the horizontal plot sizes
is not as distinct as that of main and subplots as in the split-plot design. Randomization
in strip-plot is also different from that of split-plot. Here the vertical and horizontal
factors are independently randomized on the rows and columns of the experimental area.
106
Example
Suppose that we study the properties of 4 wheat cultivates (horizontal factor) and 4
nitrogen rates (vertical factor), and the experiment is a strip-plot design with 4
replications.
1. Divide the experimental area into 4 blocks or replications(r = 4). Further divide
each block into 4 horizontal strips. Randomly assign the 4 horizontal-strip
treatments (a = 4) to each of the strips.
2. Divide each block or replication into 4 vertical strips and randomly assign the 4
vertical-strip treatments (b = 4) to vertical strips as shown in Figure 6.4.
107
Grain Yield of 4 Wheat Cultivares Grown at 4 Nitrogen fertilizers levels
C2 40 60 62 64 65
80 61 63 69 68
120 70 72 69 70
160 72 70 82 86
C3 40 75 73 72 80
77 78 77 82
120 80 82 86 83
160 84 82 84 89
C4 40 65 68 63 72
80 68 72 74 76
120 69 68 70 72
160 73 75 74 76
Degree of Freedom
108
Analysis of Variance of Wheat Cultivator X Nitrogen response
Experiment in a Strip-Plot Design
Source d.f. SS MS F
Total 63 2865.90
Cultivar X Nitrogen
(A X B) 9 237.98 26.44 3.43**
Error (c) 27 207.64 7.69
The same analysis could be done in SAS using the following program.
Proc GLM;
Class Rep variety Nitrogen;
Model yield = rep variety Rep*var Nitrogen
Nitrogen * rep Nitrogen *var;
Test H = Variety E = rep * var;
Means variety Nitrogen Variety * Nitrogen;
Run;
This design is possible when we have a third factor to be applied on the interaction plot.
That means, the interaction plots is sub-divided into split-plots to accommodate the levels
of the third factor. Hence, two factors will be applied on vertical and horizontal and the
third factor will be on the split-plot. In our strip-plot example, we applied cultivar on
vertical and Nitrogen on Horizontal plot. Let’s now consider planting date as additional
factor and apply it on the split-plot. The design for one interaction plot is shown below
109
Vertical factor = Nitrogen
Horizontal Sowing Sowing Sowing
(cultivar) date 1 date 2 date 3
The analysis could be done both in MSTATC and SAS. In MSTATC select “RCBD 3
Factor strip plots” sub-menu.
The ANOVA layout looks the following.
Source df
Rep r-1
Horizontal (cultivar) c-1
Error (a) (r-1)(a-1)
Vertical (Nitrogen) n-1
Error (b) (r-1)(n-1)
CxN (c-1)(n-1)
Error (c) (r-1)(n-1)(c-1)
Sub-plot (sowing date) s-1
CS (c-1)(s-1)
NS (n-1)(s-1)
CNS (c-1)(n-1)(s-1)
Error (d) cn (r-1)(s-1)
Proc ANOVA;
Class rep cultivar Nitrogen sowing;
Model yld = rep cultivar rep * cultivar Nitrogen Nitrogen* cultivar
Nitrogen * rep * (cultivar) sowing sowing * rep
Sowing * cultivar sowing * rep * cultivar sowing * Nitrogen
Sowing * cultivar * Nitrogen
Sowing * Nitrogen*rep (cultivar);
Note that there are 4 types of error sum of squares and if we leave test to SAS, SAS tests
all effects against the last error mean square. Therefore, it is important that we use the
TEST statement with appropriate error term declaration (e) to obtain proper test statistics.
The first test statement says that the hypothesis to be tested is declared with H = cultivar
and the error to be used for this test is given as e = cultivar * rep. All other test which
have not been declared in the program are tested using the last error mean square.
110
Remark
The ranges of experimental design used in agriculture are numerous and it is not possible
to present all of them on training manual of this type. It is always good to go through
books and literature to develop one's own caliber in designing experiments. In reality,
you are not expected to know all kinds of experimental designs. There are always people
who will help you with this problem. So, either see senior researchers or consult
Biometricians.
An experiment with two wheat variety, two fertility level and two weeding frequency was
conducted to evaluate variety-fertilizer –weed relation. Thus, the experiment constitutes
23 factorial structure (a total of 8 treatment combinations). However, there are only 4
plots available per block, which is half of what is required for complete block. Now the
question is, is it possible in the first place to continue experimentation with such
resources only? The answer is yes, it is possible to do the experiment and analyze the
data by using what is called confounding.
Obviously all treatment combinations can't go into one block (we could only
accommodate 4 treatment combination per block), therefore we have to split treatment
combination into two set. The consequence of such decision is that all effects can not be
estimated fully. The next question is which effect to ignore and how to allocate
treatments to blocks. Important steps are as follows:
Decide the least useful effect and confound it. Often higher order interactions are less
important, particularly than the main effects, and it is common practice to confound
higher order interaction. Therefore, Suppose ABC interaction is confounded. ABC
interaction is obtained from
Where effects are represented by their higher level. For example aoboco is represented as
(1), a1b1c1 = abc, aoboc1 = c, etc. Now allocation of treatment to blocks is such that all
111
effects that have an odd number of letters with abc will be allocated to one block and all
others to second block. For example, effect aobo c1 (represented by C) is allocated to say
block 1 since it has 1 letter in common with abc and effect ab which have even number of
effect with abc (which two – a and b) is allocated to second block, etc. The plan is
therefore as follows.
Plan
Block 1 2
abc ab
a ac
b bc
c (1)
Comparing this plan with effect of interaction given above (in equation 1) it is clear that
all positives go to one block and all negatives goes to another block. For example, (ab)
which have negative sign is allocated in block 2 while (a) with positive sign is allocated
in block1. Therefore AxBxC interaction is completely confounded with block effect.
That means when we try to estimate interaction effect we are also estimating block
effect. It is practically not possible to separate the two effects. That is why we say block
and interaction effect are confounded.
Now, the next question is can we estimate other effects? The answer is YES.
For example consider. A (effect) = (abc +ab+ac+a) – [(1)-b-c-bc)]. Now, in the first
block we have two (abc and a) positive and two (b and c) negative effects. Similarly, in
the other block we have two from each sign. Hence effect of A is
112
1983).
Rep I Rep II
p k d npk npk d p dnk
Block a 45 55 53 36 Block a 43 42 39 34
dnk dnp dpk n n dnp k dpk
41 48 55 42 47 52 50 44
dp nk dk pk nk Dp (1) np
Block b 50 44 43 51 Block b 43 52 57 39
dnpk (1) dn np pk dk dnpk dn
44 58 41 50 56 52 54 42
Step 1. Calculate total for each treatment combinations, the block and replicate total, and
the grand total.
Step2. Calculate factorial effect totals. The treatment combinations and their total yields
are placed in a systematic order in the first two columns, and the remaining
columns give the factorial effects. The factorial effects total are shown in the
bottom of table below. For example, the factorial effect for D is calculated as
-115 + 95 – 89 + …….+98 = -8
113
TABLE 6.2 CALCULATION OF FACTORIAL EFECT TOTALS IN A 24 EXPERIMENT
P 84 - - + + - - + - + + - - + + -
Dp 102 + - - + + - - - - + + - - + +
Np 89 - + - + - + - - + - + - + - +
Dnp 100 + + + + + + + - - - - - - - -
K 105 - - + - + + - + - - + - + + -
Dk 95 + - - - - + + + + - - - - + +
Nk 87 - + - - + - + + - + - - + - +
Dnk 75 + + + - - - - + + + + - - - -
Pk 107 - - + + - - + + - - + + - - +
Dpk 99 + - - + + - - + + - - + + - -
Npk 79 - + - + - + - + - + - + - + -
Dnpk 98 + + + + + + + + + + + + + + +
1502 -8 -120 32 14 88 50 8 -12 -14 -32 18 28 -22 -32 50
The total for DNPK is not required, since owing to the confounding DNPK will not appear explicitly in the
ANOVA.
Step 3. The contribution of each factorial effect to the treatments SS in the analysis of
variance is now obtained. Since there are 32 plots, the square of each effect is divided by
32. The ANOVA is shown below. For example, to find sum of square (ss) for D effect,
square factorial effect of D(-8) and divide by 32. Hence SS for D = (-8)2 = 64 = 2
32 32
Source d.f SS MS
Replication 1 3.1
Blocks in reps 2 123.2 61.6
D 1 2.0
N 1 325.1
P 1 6.1
K 1 4.5
DN 1 32.0
DP 1 242.0
NP 1 78.1
DK 1 6.1
NK 1 32.0
PK 1 24.5
DNP 1 2.0
DNK 1 10.1
DPK 1 15.1
NPK 1 32.0
Error 14 340.0 24.29
Total 1277.9
114
Note that DNPK is omitted from ANOVA since its SS is included in the blocks SS.
Proc ANOVA;
Class rep block D N P K;
model yield = rep block (rep) D N P K D*N D*P N*P D*K N*K P*K
D*N*P D*N*K D*P*K N*P*K;D*N*P*k;
Means D N P K D*N D*P N*P D*K N*K P*K D*N*P
D*N*K* N* P*K D*N*P*K;
Run;
Error 14 339.75000000
115
Partial confounding
Example. Consider 23 factorial in 3 reps each with two blocks of size 4. Different
effects are confounded in different replications.
AC (effect) = {abc +ac + bc +(1)} - {ab +bc +a +c}
BC (effect) = {abc + bc + a +(1)} - {ab-ac-b-c}
Replication
Rep I Rep II Rep III
1 2 1 2 1 2
Abc Ab Abc Ab Abc ab
A Ac ac Bc Bc ac
B Bc B A A b
c (1) (1) C (1) c
confounded ABC AC BC
ANOVA
Source df
Blocks 5
A 1
B 1
C 1
AB 1
AC 11
BC 11
ABC 11
Error 11
Total 23
116
7. Cluster Analysis for Characterization and
Classification
7.1 Introduction
Cluster Analysis is a multivariate statistical analysis technique for grouping observation
(entries or environments) into classes based on the characteristics measured prior to the
analysis. For example, classify entries based on their site mean. Elements within the
same group or cluster are relatively homogenous and element between clusters are
relatively heterogeneous. This technique was initially developed and used in the area of
taxonomy for several decades and eventually introduced to the breeding and agronomy
areas in early 70's.
Example 1: Taxonomy
Let’s assume that we have m number of accessions and we have measured n number of
traits for each. Further assume that the experiment was done in RCBD with 3
replications. That means we have 3m number of observations in the whole data set.
However for cluster analysis replications are not needed because cluster classifies
accessions based on their respective mean traits. Therefore, we need to take average over
replication and prepare data with accessions as cases (row) and traits as variables
(column).
Traits
1 2 3 4 . . . n
Accessions (plant (hair (seed)
length) color) size
1 . . . . . . . .
2 . . . . . . . .
3 . . . . . . . .
. . . . . . . . .
. . . . . . . . .
m . . . . . . . .
117
Example 2: plant breeding
Cluster analysis is also used to group genotypes into homogenous sets based on their
response to the environments considered. This result normally complements the stability
parameters calculated using the regression techniques we described earlier. Therefore,
clustering is used for characterization, classification and study of G x E interaction to
help in variety development programs.
In technology adoption studies where a number of farmers are asked opinion about a
particular variety, clustering may be used by grouping the respondents based on their
opinions so that differences in response about the adopted technology is fairly evaluated.
In developing new varieties, plant breeders are always faced with the choice of two
strategies:
(II) to breed for varieties each adapted to a specific sub-region within the region.
118
obvious that for the second option, contrasting sub-regions within which environments
are similar need to be identified. However, even for the first option, representative
environments from contrasting sub-regions are also required to effectively screen for
wide adaptability during the intermediate and final stages of testing. Various approaches
were suggested to stratify environment into sub-groups, but so far, cluster analysis based
on grain yield and other characteristics has been widely used.
Cluster analysis has been used to classify genotypes. Based on similarity of response to
environments, genotypes can be assigned into qualitative homogeneous groups. This
information complements the stability parameters obtained from the regression analysis
made popular by Finaly & wilkinson and Eberhart & Russell.
119
CHOICE I
Nonexclusive Exclusive
CHOICE 2
EXtrinsic Intrinsic
CHOICE 3
Hierarechical Nonhierarchical
CHOICE 5 CHOICE 4
CHOICE 6
Monothetic Polythetic
All exclusive intrinsic classifications are used for the same sort of purpose, so the strategies
within the general class are largely interchangeable and choice will largely be determined by
pragmatic consideration, mainly of a computational nature. The most important choice is that
between hierarchical and non-hierarchical strategies. A hierarchical strategy always optimizes a
route between the entire population and the sets of individuals of which it is composed. The rout
may be defined by progressive fusion of divisions.
With a non-hierarchical strategy, it is the structure of the individual groups which is optimized, since these are made
as homogeneous as possible. For cases in which homogeneity of groups is of prime importance, the non-hierarchical
strategies are very attractive, however their current state of development lags far behind the hierarchical strategies.
An agglomarative strategy is one that proceeds by progressive fusion, beginning with the
individuals and ending with the complete population. All agglomerative classifications are
polythetic. This is the most commonly used strategy.
A polythetic system is one based on a measure of similarity applied over all attributes, so
that an individual is grouped with those individuals which, on the average, it most
resembles. It appears that the ideal hierarchical system is divisive polythetic, but only a
limited number of programs are available.
√[(X11-X12)2 + (X21-X22)2]
Usually, there is no need to take the square root since the operation does not change the order of
how close the points are. Therefore, the alternative approach is known as squared Euclidean
distance. When there is a number of variables, the presentation on paper becomes impossible but
can only be presented as a formula. If there are s variables, the squared Euclidean distance
between the ith and jth entry can be given by
When the variables have different units, one needs to standardize the data before computing the
distance. This is known as Standardized Euclidean Distance. One can use the range for this
purpose, but commonly we use the standard error, i.e., subtract the variable mean from each
observation and then divide by the phenotypic standard error of that variable. After
standardization, each variable will have a mean of zero, unit variance and become dimensionless
(no units of measurement). If the term standardized Euclidean distance is used without further
explanation, it is this particular form of standardization that is implied.
Another common measure of similarity is the simple correlation coefficient. It can be shown
mathematically that the standardized squared Euclidean distance between two variables is related
to the correlation coefficient, and thus they provide identical information
For grouping environments based on grain yield of a number of genotypes, Euclidean distance
should also be standardized when differences in environmental mean yield are large due to
recognizable reason. Otherwise, there will be a tendency for groups having similar yields to
fuse together (Fig.3). It should be noted that high and low yielding environments which rank
cultivars identically are more similar in a selection sense than 2 environments of similar yield
which rank cultivars differently.
121
The distance between 2 clusters taken as the distance between the 2 nearest points, one from
each cluster .The strategy is called space-contracting; as a group grows it becomes progressively
easier to join. Clustering is very weak or non-existent. Individuals tend to join a group one
after another, a phenomenon known as chaining. As such, it is undesirable for most biological
applications.
The distance between 2 clusters taken as the distance between the 2 furthest points, one from
each cluster. The strategy is called space-dilating; it sharpens any discontinuities so that
clustering is intensified. The data are thus distorted. It is possible that outliers may be swept into
a 'non-conformist group' whose members share only the property that they are rather unlike
everybody else, including each other. Such strategy is useful when it is known or suspected that
no sharp discontinuities exit and it is wished to produce as 'clean' a classification as possible,
i.e./ to make the best of a bad job.
3. Centroid:
The distance between 2 clusters is defined as the distance between the group centroids. The
centroid is the point whose coordinates are the means of all the observations in the cluster. If the
cluster has one observation, then the centroid is the observation itself. It is not monotonic.ie,
later fusions may form at a smaller distance. When drawn on paper, there will be 'reversals'
which cause a later fusion to lie below earlier ones and make the dendrogram difficult to follow.
4. Group-average:
The distance between 2 clusters is defined as the average distance between all pairs of
observations from each cluster. It is nearly space-conserving and can be regarded as such for
most practical purposes. It is monotonic, without the problems of centroid linkage. It avoids the
extremes of either the single or maximum linkage methods.
Several simulation studies comparing various methods of cluster analysis has been performed.
Although the results of such study were inconsistent and confusing in many respects, average
linkage or ward’s minimum variance found to show best performance in grouping cases.
It should therefore be noted that different fusion strategies produce quite different results, so the
choice of an appropriate fusion strategy is extremely important.
122
7
3
2
1
b c a d g e f
For the above figure, beginners can easily jump to the conclusion that b and f are the most
unlike, simply because they are at opposite ends. However, we should not forget that a
node simply represents a fusion between two equivalent entities, and any one of the nodes
can be rotated without altering the sequence of fusions (see the figure below)
7
6
5
Fig. The same dendrogram as
4 Fig.13.1; the original
dendrogram has been
3 rotated round two of its nodes.
2
1
d b c a e f g
The final partition or the population is defined by cutting the dendrogram. Choosing the level of
the cut, and thus the number of classes needs a lot of judgment, except when real well defined
groups exit. As a general role, large increases in distance between two sequences should be a
signal for examining those particular steps. A large distance indicates that at least two very
different observation exist, one from each of the clusters just combined. The result should be
inspected from a subject matter perspective.
In cluster analysis determining the number of clusters is most difficult job. Because, there are no
satisfactory methods for determining the number of population clusters for any type of cluster
analysis. Several attempts has been done through simulation studies to search for best methods
of determining number of classes. A notable attempt has been done by Milligan and Cooper
123
(1985) and Cooper and Milligan (1984). They compared thirty methods for estimating the
number of population clusters using four hierarchical clustering methods. They identified three
best performing criteria in their simulation studies, namely, pseudo F-statistics, pseudo t2 statistic
and the cubic clustering criteria (ccc). The procedure for selecting appropriate number of clusters
is to look for consensus among the three statistics. That is local peaks of the ccc and pseudo f
statistics combined with a small value of the pseudo t2 statistics and large pseudo t2 for the next
cluster fusion.
Basically, this criteria locates all feasible cluster numbers and some cluster numbers can be
equally appropriate. For example, the criteria may suggest a 2 or 4 clusters from a particular data
set, but judgement should also consider the biological aspect of the data. Therefore, once
candidates are selected by cluster analysis practical aspect of the research result may be
considered for final decisions. Cluster analysis uses different clustering methods, and these
methods use different measures of distance and fusion strategies which can lead to very different
grouping. It is therefore important to make several analyses using different methods on the same
data to help in final decisions.
Part of the data was taken from pulse program to illustrate cluster analysis in three packages.
The SAS package (version 6.12) was used to produce the following.
The data set contains 50 entries and for each entry 15 traits were measured. The traits include
plant height, days to heading, days to Maturity, number of tillers, spikes, kernel weight, grain
yield and other quantitative characteristics.
SAS program used to generate output below is:
Three clustering methods were used: average linkage, centroid and ward’s minimum variance. In
all of them the pseudo F reached it peak at cluster three. The pseudo t2 reached its minimum at
cluster 3 in both average linkage and centroid clustering with higher values below and above it.
124
Therefore, it seems that all the three methods agree on same number of clusters. Looking for
second candidates is a bit difficult because if we go for cluster2, the pseudo F is smaller than that
of cluster 3 and t2 is larger than that of cluster 3. This particular data seems to have a clear
pattern where entries could be ambiguously grouped. Hence we may judge that the entries could
be classified in to 3 clusters based on 15 characteristics.
The next step in the SAS program is to go back to line 10 and write n clusters = 3 and re – run the
program so that we obtain list of entries that fall in each group. Old version of SAS can also give
dendrogram but it is not presentable for reporting purposes. The new version has shown
tremendous improvement in this regard.
49 6 7 2 1012.01 . 0.031760
48 9 10 2 317.98 . 0.074484
47 64 63 2 218.43 . 0.089831
46 14 15 2 179.83 . 0.097891 T
45 24 23 2 157.61 . 0.105336 T
44 59 60 2 147.21 . 0.105336
43 1 4 2 131.48 . 0.125039
42 CL46 12 3 119.28 1.88 0.126043
41 3 2 2 114.43 . 0.127040
40 25 CL45 3 108.69 1.73 0.130950 T
39 54 58 2 106.78 . 0.130950
38 66 65 2 103.78 . 0.142035
37 CL42 16 4 99.03 1.84 0.146693
36 CL49 5 3 94.16 29.00 0.148967
35 70 69 2 91.54 . 0.166551
34 57 56 2 89.24 . 0.172501
33 67 62 2 87.70 . 0.173956
32 19 20 2 86.75 . 0.175400
31 61 68 2 85.67 . 0.183824
30 CL40 21 4 82.86 2.97 0.186997
29 CL48 CL43 4 78.20 5.76 0.189232
28 26 27 2 78.59 . 0.193188
27 30 29 2 78.99 . 0.199608
26 CL34 51 3 79.19 1.49 0.201494
25 22 28 2 80.27 . 0.202119
24 52 CL44 3 79.55 5.30 0.216574
23 CL38 CL31 4 78.01 2.68 0.222886
22 CL36 CL29 7 73.11 5.28 0.223451
21 CL37 11 5 73.23 4.18 0.229574
20 17 18 2 75.38 . 0.232304
19 CL28 CL25 4 76.44 1.90 0.238200
18 CL26 53 4 77.56 2.20 0.258019
17 CL39 CL24 5 75.75 4.74 0.270737
16 CL21 CL32 7 72.52 5.92 0.292554
15 CL47 CL33 4 72.18 8.11 0.295385
14 CL23 55 5 74.11 2.96 0.306076
13 CL27 CL19 6 73.44 4.35 0.327951
12 CL22 CL41 9 71.27 8.24 0.337719
11 CL16 CL20 9 71.83 4.54 0.347861
10 CL30 CL13 10 69.41 7.14 0.362641
9 CL15 CL17 9 66.70 9.16 0.412084
8 CL35 CL14 7 67.20 9.34 0.462813
7 CL8 CL9 16 59.58 9.87 0.518391
6 CL12 8 10 66.63 6.77 0.531920
5 CL6 CL10 20 55.06 24.66 0.595437
4 CL7 CL18 20 56.60 12.12 0.668186
3 CL11 13 10 77.96 11.03 0.734190
2 CL5 CL3 30 71.74 38.35 0.901291
1 CL2 CL4 50 . 71.74 1.262827
125
Centroid Hierarchical Cluster Analysis
49 6 7 2 1012.01 . 0.031760
48 9 10 2 317.98 . 0.074484
47 64 63 2 218.43 . 0.089831
46 14 15 2 179.83 . 0.097891 T
45 24 23 2 157.61 . 0.105336 T
44 59 60 2 147.21 . 0.105336
43 CL46 12 3 126.67 1.88 0.116152
42 25 CL45 3 114.23 1.73 0.119891
41 1 4 2 110.93 . 0.125039
40 3 2 2 108.69 . 0.127040
39 CL43 16 4 100.51 1.84 0.130091
38 54 58 2 100.35 . 0.130950
37 66 65 2 99.03 . 0.142035
36 CL49 5 3 94.16 29.00 0.148118
35 70 69 2 91.54 . 0.166551
34 57 56 2 89.24 . 0.172501
33 CL42 21 4 83.53 2.97 0.172987
32 67 62 2 83.12 . 0.173956
31 CL48 CL41 4 76.77 5.76 0.174680
30 19 20 2 77.59 . 0.175400
29 CL37 68 3 76.55 2.17 0.181060
28 CL34 51 3 76.02 1.49 0.182101
27 CL36 CL31 7 67.02 5.28 0.186040
26 26 27 2 68.66 . 0.193188
25 29 CL26 3 69.29 1.34 0.193840 T
24 22 28 2 71.13 . 0.202119
23 CL29 61 4 71.31 1.96 0.204051
22 52 CL44 3 72.25 5.30 0.210072
21 CL39 11 5 72.44 4.18 0.214673
20 17 18 2 74.60 . 0.232304
19 CL28 53 4 75.21 2.20 0.232906
18 30 CL25 4 76.33 1.88 0.234226
17 CL38 CL22 5 74.68 4.74 0.239490
16 CL21 CL30 7 71.62 5.92 0.255396
15 CL18 CL24 6 69.79 3.63 0.267791
14 CL47 CL32 4 70.58 8.11 0.278692
13 CL23 55 5 73.44 2.96 0.279088
12 CL16 CL20 9 72.37 4.54 0.288023
11 CL33 CL15 10 68.00 7.14 0.297711
10 CL27 CL40 9 69.41 8.24 0.305794
9 CL14 CL17 9 66.70 9.16 0.350901
8 CL35 CL13 7 67.20 9.34 0.426808
7 CL8 CL9 16 59.58 9.87 0.397489
6 CL10 8 10 66.63 6.77 0.503030
5 CL6 CL11 20 55.06 24.66 0.509651
4 CL7 CL19 20 56.60 12.12 0.577081
3 CL12 13 10 77.96 11.03 0.708941
2 CL5 CL3 30 71.74 38.35 0.788646
1 CL2 CL4 50 . 71.74 1.105995
126
Ward's Minimum Variance Cluster Analysis
T
Pseudo Pseudo i
NCL Clusters Joined FREQ SPRSQ RSQ F t**2 e
127
---------------------------------- CLUSTER=1 ------------------------------
X1 X2
6 1
7 2
9 3
10 4
24 *
23 *
1 5
4 8
3 9
2 *
25 *
5 6
26 *
27 *
30 *
29 *
22 *
28 *
21 *
8 7
X1 X2
64 *
63 *
59 *
60 *
54 *
58 *
66 *
65 *
70 *
69 *
57 *
56 *
67 *
62 *
61 *
68 *
51 *
53 *
52 *
55 *
X1 X2
14 *
15 *
12 *
16 *
19 *
20 *
17 *
18 *
11 *
13 *
128
Cluster analysis can also be done using SPSS and CLUSTAN GRAPHICS (a small
program written by Dr. Wishart, dedicated to cluster analysis). SPSS version 10.1
produces a number of outputs. It produces agglomeration schedule which shows history
of how clusters combined, cluster membership and Dendrogram for average linkage
clustering. To get cluster membership you must supply number of clusters that you think
is appropriate in a sub-menu called “statistics” of the cluster dialog box. Thus SPSS and
SAS follow very similar procedure in their outputs, but SAS gives definite criteria for
judging the number of appropriate clusters. Denderogram produced by SPSS is attached
for comparison.
CLUSTAN GRAPHICS produces only Dendrogram and cluster means for each
characteristics measured. The difference between this package and others is that clustan
automatically determine number of clusters appropriate for a particular data set, and
highlight the clusters with colours. This might have advantages or disadvantages. The
advantage is that users will not have many problems to identify appropriate number of
clusters. The disadvantage is that users do not have alternatives to consider biological
aspect and choose one among given alternatives. Dendrogram from CLUSTAN is also
included for references.
129
SPSS dendrogram
130
8. Establishing Field Experiments
Establishing field experiments is one of the most important area requiring due attention
before beginning a trial. This is also an appropriate stage where homogeneity of the land
made available for an experiment and the number of treatments selected for this purpose
are examined. Particularly, the trade off between cost and precision or availability of
resources and level of precision that can be attained with such resource should be
thoroughly examined prior to execution.
Is there sufficient information on land & soil characteristics so that the history of land is
well known to set up local control mechanisms like blocking structure and direction and
possibility for use of covariate?
Is the soil uniform (texture, depth . . . .)? The soil need to be uniform to make the
blocking structure efficient, otherwise there is a possibility that genotype by environment
interaction component, which hinder yield estimation, appear in the analysis. It is
therefore advantageous to know past cultural practices and ensure adequate access to
roads for proper follow up of the experiment.
♦ Avoid previous experimental areas where treatments may have residual effects on
the soil, since they might create differential response of varieties within the same
blocking structure.
♦ Avoid areas used as unplanted alleys in previous works.
♦ The choice of proper experimental design that fit the pattern of soil heterogeneity is important. There
are now a number of possibilities in this regard. One obvious alternative is, first to observe pattern of
soil heterogeneity and select design that can match the pattern. Recent development in this area, also
known as spatial statistics, is becoming a powerful tool that can detect spatial variation and the
information is used during analysis.
♦ Choosing proper plot size and shape is also a serious problem in several research
programs because plot size and shape being used in several commodities are based
on either experiences of other countries or own personal Judgement.
♦ Orientation of plots & blocks should be carefully done as they have to go with the
pattern of soil heterogeneity.
131
♦ Increase replication. This may hold to obtain more information and estimate
experimental error more precisely.
8.3 Randomization
Perform randomization that suits your experimental design. Note that each experimental
design have their respective randomization. For example, in RCBD randomization is
done within block. In split-plot two stage randomization is required. First randomize
main plot factor on the main-plots per block, then randomize split –plot factor on split-
plot per main-plot. Particularly, use separate randomization for each location if multi-
location trail is thought. Otherwise same treatments placed side by side in all blocks and
problem of serial correlation may occur. Therefore, do not send same copy of
randomization plan to cooperating centers.
132
Factor variables Characteristics
Plot Location Rep Variety Nitrogen Row Column Plant Seed/ Seed Yield/of
No height pod size
1 1 1 1 1 1 1 . . . .
1 1 1 1 2 . . . . . .
1 1 1 2 1 . . . . . .
1 . 1 2 2 . . . . . .
. . 2 . . . . . . . .
. . 2 . . . . . . . .
. . 2 . . . . . . . .
. . 2 . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . 4 . . . . . . . .
. . 4 . . . . . . . .
. . 4 . . . . . . . .
. 1 4 . . . . . . . .
. 2 1 . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
. . . . . . . . . . .
2 4 2 2 . . . . . .
Plot number could be important for some packages and not for others, therefore, it is
good idea to keep it in the structure. Row and column are also included here even if the
design was not row-colum design. This is because, recent developments in methodology
enabled us to do what is called “nearest neighbor analysis” which takes in to account the
spatial variability available in the experimental field and make adjustment to parameters
estimated from the data.
133
9. Collection of Experimental Data and
Underling Assumptions
9.1 Data Collection
We have already discussed in previous section design protocol, and now we assume that
data to be collected has been exhaustively listed in material and methods and the
researcher is already aware of what data to be collected. The concern here is therefore
techniques for collecting data from the field. Techniques for collection data depends on
the type of information required. For measuring grain yield, for example, we use
destructive method. That is harvesting the whole plot (for standard on station plots). For
large plots and for some plant characters, we normally sample plants, group of plants or
part of a plant (e.g. leaf) to take a data. This is basically identifiable in the beginning and
should be indicated in materials and methods.
1) Sampling unit
2) Sample size
3) Selection method
Sampling is required because, we can't measure all plants in a plot, e.g., leaf area, plant
height, panicle weight etc. Therefore one should plan to achieve fair representation of
samples by employing a reasonable sampling methodology. Use same sample plant at
all stages of observation for measuring the same character. Otherwise there is no sense
of repeated measurement and the values obtained may not be comparable as they are.
Use proper measurement technique and record data in the proper manner. Take care in
selecting unit of measurement (e.g meter) and make sure to decide on number of decimal
places. Decide what data to be collected at the planning stage of the trial and don’t
gather data just because you get them. The danger of collecting too much data is (i)
creating confusion because it might be difficult at later stage which one to choose and
use (ii) if you can’t use all the data collected then it turn out to be waste of resource and
time.
Date of recording data must be specified for some characteristics (e.g. days after
seedling). For qualitative data, codes should be consistently used. Care should also be
taken regarding definition of measurements. For example, plant height for seedling or
Juvenile plants is the distance from ground level to the top of the tallest leaf. For mature
plants, it is the distance from ground to tallest panicle. Grain weight is a weight of
cleaned and dried grain harvested from a unit area. Moisture content should also be
determined abruptly. Leaf area could be measured in different ways and it is subject to
your own ability to manage different methods, because leaf of different plants have
different shapes which can’t be generalized using single geometry.
134
Exceptionally large values should attract your attention, because they could be outliers.
If a value is false observation, then it will have a due influence on the ANOVA or
regression results, often misleading conclusions.
Often you may design your experiment very well, say by using blocking or otherwise to
adjust for systematic patterns in the field, but there could be unexpected damage (e.g by
rodents.) in some of the plots. You may register those plots that have such problem and
calculate an index for them based on severity and then apply what is called covariance
analysis to adjust for the damage (see 4.5 for detail).
The last activity in this regard is to prepare data for computer. The structure of the data is
nearly similar for all statistical packages (the structure is given in 7.4). Plots are treated
as case numbers (rows) and the characteristics measured (known as variables) occupy
columns. These days there are ample facilities to transfer data file from one package to
another and one may use any statistical package, spreadsheet or database for data entry
since it is possible to convert data to different format when required.
1) Assumption of Normality
Analysis of variance and several test procedures (f, t, z) require data to be approximately
normally distributed. In agriculture or biological sciences, nature has endowed creatures
with normal distribution and there is no serious problem in this regard. Often, such
problem arises for count, scaled or proportion data.
135
♦ Simple plot of data will roughly help to see the approximate distribution of the
data.
When means from different population are to be compared, one important assumption is
that populations from which the means are obtained roughly have similar variance. If
variance of different population is very different, it is not possible to pool variance and
perform significance test. This is important assumption to be checked when working
with multi-location and several year trial and when a combined analysis of variance is
required.
Check in particular if
Method of detection
1) Use F-ratio of the two variances. If the ratio is much greater than 2, then the
two population belong to different level of variability.
2) Use X2 test (compare the ratio of variances with critical points of X2)
For this example homogeneity is suspicious because the treatment means are very
different. Hence, the s.e of 6.4 underestimate variability for control and over estimate
that of insecticide C.
Additivity: -
For RCBD, the difference in yield between the treatment means should, on average, be
the same in each block (i.e. no block * treatment interaction).
136
Example (Fictious data)
This is often a source of controversy in most research review meetings because no one
knows what size of a block is exactly uniform (optimum) to avoid problem of non-
additivity. Obviously, as number of treatments to be included in a complete block
increase, block loosing homogeneity. This gives rise to occurs of block by treatment
interaction which means that since plots in the block are not uniform some treatments are
favored. Therefore, it is not valid to compare treatments under such condition. The
solution to such problems is to determine optimum plot size for different crops so that
optimum block size are consequently fixed.
Although means appear different, they are not significantly different in the above table, when subjected to
analysis of variance, because data is multiplicative. So transforming data using Log is suggested.
After transformation the F value was found to be 24 which shows a highly significant
difference between the treatments. Therefore, this is evidence that failure of assumption
creates illusion on conclusions.
4. Independence
137
Two Special Cases of Assumption Failure
1) Poisson
2) Binomial
Presentation
Presentation is another problem area if assumption fail and a remedy is adopted. The
reason is analysis and conclusions are appropriate on the transformed scale, but
presentation of the means on transformed scale does not give sense to users. So, there
are two group of opinions. One group provokes the idea of presenting raw means only
while inferring on transformed data. The other group provokes the idea of presenting
both transformed and raw mean hoping that readers would also understand, by looking
the two scales side by side, why transformation was required in the first place.
138
10. Statistical Analysis In Different Packages
10.1 Overview
There are several statistical packages available. However, few of them (SAS, GENSTAT
and SPSS) have totally monopolized the market. SAS is the major one in terms of the
number of modules (module is a program devoted to specific task like regression analysis,
etc…), number of users and the variety of model incorporated. It also covers a wide
range of disciplines like statistical/econometrics, biomertical, bio-medical, industrial,
socio-economics, operations research, quality control, GIS and mathematical modeling.
For this reason it is being used by several institutions all over the world. Such diversity in
analytical capability pushed price of this package so high that it is becoming unaffordable
to the third world.
SPSS(Statistical package for the social science), American origin, was initially developed
targeting the social sciences where survey work is most common. These days, however,
they are trying to widen their horizon for marketing and are in the process of
incorporating models and designs from other areas. Particularly this last version of SPSS
(10.1) included Generalized models, multi-variate analysis and others. However, SPSS is
not preferable for designed experiments since there is no design principle in this package.
There are also other packages like, MSTATC MINTAB, INSTAT, AGROBASE,
SYSTAT, STATISTICA, CLUSTAN GRAPHICS and others which are limited in
coverage and have smaller data management capacity. Most of them were written for
specific purposes. MSTATC, for example, was written to handle agronomic and breeding
experiment that are designed in complete block or lattice designs, and it can’t do analysis
for trials with incomplete block structure.
As described earlier, SAS is quite ok on most activities like data management, analysis of
both designed & non-designed experiments and surveys. As is evident in table below,
most statistical packages have data management problems, most of them allow users to
enter data into limited spreadsheet structure pre- determined and are not flexible to
change format or data values or structure. In all such packages, data file merging,
concatenating, transposing, rearranging variables, grouping cases are difficult. In SAS
and SPSS this activities can be easily done.
139
10.2. AGROBASE versus MSTATC
They are:
1.
Data/ file management facilities,
2.
Randomization and layout,
3.
Analysis and
4.
Experimental procedures/management. It also offers other small facilities
like programming in AGROBASE.
The first screen of AGROBASE (main window) is composed of three features: the
AGROBASE MAIN MENU, BUTTON BAR. And AGROBASE system status.
The basic difference between AGROBASE and MSTATC is that the former has better
data management facilities, advanced statistical models and designs, randomization
procedures, pedigree management for some crops and other useful utilities which are not
available in the MSTATC.
The SAS system is a comprehensive and flexible information delivery system that
supports of
By taking advantage of features built into windows, the SAS system has increased power
and flexibility so that it can communicate and share data with other windows applications
140
(like EXEL, ACCESS …)
SAS system also has a menu driven way, SAS/Assist, which allows us to carry out most
data operations and analysis by mouse selections rather than by writing programs.
10.4. Genstat
Genstat is a comprehensive statistical system designed to summarize and analyze data on
computer. It provides standard statistical analysis by use of menu system, which can
itself be modified and extended through the use of command Language. It manages data
that entered within Genstat own spreadsheet or imported from other file format. It
performs standard tests, summarizes, compares, fit distribution and graphs. It transform
data using calculation facilities inbuilt. It analyzes data generated from one-way analysis
of variance to complex multi-strata designs. It assess relationship between variables
using regression and correlation, perform multivariate analysis such as principal
component and cluster analysis. It can even go far to time series and spectral analysis.
For advanced statistical methods, Genstat furnishes programming Language which, if
managed properly, can do most of the modeling practices available in agricultural
research.
1. The major statistical packages available on the market may be compared as follows,
Major Statistical packages
Data Non –
Soft ware type Management Designed exp. Surveys Designed
SAS E E E E
GENSTAT G E F F
MINTAB P G P F
SPSS E P E G
SYSTAT F G F F
INSTAT P G P F
AGROBASE F G P P
MSTATC F F P P
KEY:
E = Excellent
G = Good
F = Fair
P = Poor
141
2. Output of different packages for same data set
The following ANOVA tables has been given in the same way they would have been
printed by their respective packages only the ANOVA component is presented here for
saving space, otherwise most of these packages also produce other output like means,
standard errors and CV.
ANOVA of MSTATC
ANOVA OF MINTAB
Source DF SS MS F F
Rep 3 248.550 82.85 2.89 0.0792
Treatment 4 1068.3 267.075 9.32 0.0012
Error 12 343.7 28.6417
Total 19 1660.55
MTB >
ANOVA tables for MSTAT and MINTAB have similar structure. They are basically
developed for designed experiment and understands what rep and treatment means.
MINTAB is a small program but can handle most basic designs and analysis. It manages
incomplete blocks of modest non-orthogonality, fractional factorial generation and
composite designs. Among multivariate methods, MINTAB handles principal
component, dicriminant analysis and others. Unfortunately, cluster analysis is not
available in release 9.2 or earlier. It can read ASCII data and export it. SYSTAT is a bit
different because it considers rep* treatment interaction. SPSS does not suit for
designed experiments because, the ANOVA table is slightly different. For example, it
considers all the factors that entered into the model as treatments and produces
interaction of all terms if the data is submitted directly. For this example, “univariate”
sub menu under “General linear models” menu were used in SPSS 10.1 with unique sum
of square. To avoid interactions of rep by treatment one must select “model” option and
142
use “custom” declaration of terms and select only main effects so that error could be
calculated. SPSS do not produce outputs like CV, grand mean and LSD. It can only
analyze designs of simple factorial nature. In addition, the way it calculate degree of
freedom for interaction is odd and in principle not correct for designed experiments.
Here rep, and treat have 3 and 4 df respectively which will be 12 df for their interaction,
however SPSS allocated only 11 and assigned 1 df for error. Similarly, SYSTAT did not
produce the error, rather it put the 12df as rep* treat interaction just like SPSS.
The recent version (version 10.1) of SPSS produce more structured and more general
ANOVA output. It now operates under the “General linear model” menu, as opposed to
“ANOVA model” in earlier versions, and give a general ANOVA structure, which
serves for a variety of purposes. For example, it furnishes for principle of random and
fixed effect models. A factor is termed as fixed if the level of that factor are the only
available once in the whole of population and the levels were not selected. Random
refers to the fact that levels of factors has been chosen from population which have other
similar factor levels and this accounts for the random variation in selections. This
principle is now used in both SAS and GENSTAT. In the model sub-menu type of sum
of square must be chosen along with whether interaction should included. SPSS uses
four types of sum of square just like SAS. As mentioned earlier you must decide how to
get error sum of square, particularly in several factor- factorial experiment where SPSS
produces interaction of rep with all other factor which should normally pooled and
termed error in other packages. One of the quality of SPSS is that it can handle
incomplete factorial combinations and adjust treatment means. But, it can not handle
most experimental designs used in breeding or agronomy.
On the other hand, GENSTAT produce a general ANOVA structure which can serve for
all types of designs. It designates stratum to each level of random variation. Here, the
REP stratum has only the rep, while treatment enters the next stratum. Also note that
ANOVA results for most packages are the same except for rounding error, but that of
SPSS is completely different from the rest. Therefore it should be noted that SPSS may
not be appropriate for designed experiments.
ANOVA OF GENSTAT
143
***Standard errors of differences of means***
Table TREAT
Rep 4
D.f. 12
s.e.d. 1.894
ANOVA OF SYSTAT
Rep (4 levels)
1. 2 3, 4
TREAT (5 levels) 2, 3, 4, 5
Analysis of Variance
Source Some-of -Squares DF Mean-Square F. Ratio p
REP 248.55 3 2.89 . .
TREAT 1068.30 4 9.32 . .
REP*TREAT 343.7000 12 28.6417 . .
ERROR 0.0 0 . .
ANOVA OF SPSS
14 Aug 88 SPSS for WINDOWS Release 6.1
144
ANOVA of SAS
REP 4 1 2 3 4
FACT 5 1 2 3 4 5
Error 12 343.70000000
145
Combined analysis of 2 Factors in Mstatc
(Assuming RCDB)
146
Combined Analysis of same data set in SPSS
* * * ANALYSIS OF VARIANCE * * *
YIELD
By LOC
REP
FACTA
FACTB
147
11. Regression and Correlation Analysis
In previous section the objective of analysis was to either see whether there is differences in
mean values between groups (such as varieties or treatments) or to classify observations
(landraces, species, are etc...). But in this section we will study relationship between two or
more variables.
Response to treatments can be exhibited either by the crop or livestock in terms of changes in
such biological features as grain yield or animal weight or by surrounding environment in
terms of changes in such features as insect incidence in an entomological trial and soil
nutrient in a fertility trial.
To assess association between traits or characters there is a need for statistical procedure that
can simultaneously handle several variables. If two plant characters are measured to
represent crop response, the ANOVA and mean comparison procedures can evaluate only
one character at a time, even though response in one character may affect the other, or
treatment effects may influence both characters. Regression and correlation analysis thus
allows a researcher to examine any one or a combination of the three types of associations as
long as the variables considered are measured quantitatively for theoretical reasons.
Regression analysis describes the effect of one or more variables, called independent
variables, on a single variable, called dependent variable. The latter is expressed as a
function of the former.
Correlation analysis, on the other hand, provides measure of the degree of association
between two or more variables or the goodness of fit of a prescribed relationship to the data
at hand.
148
Linear Relationship
The functional form of the linear relationship between a dependent variable Y and
independent variable X is represented by the equation:
Y= α + βx
When there is more than one independent variable, say K independent variables (x1, x2,....,
xk), the simple linear functional form of the equation Y = α + βx can be extended to the
multiple linear functional form
The presence of βi (βi # 0) indicates the dependence of Y on Xi. Thus, the test of
significance of each βi to determine whether or not βi = 0 is an essential part of the
regression analysis.
Ho: β = O, i.e there is no relationship between dependent (y) and independent (x).
Ha: β # O, i.e there is linear relationship between the two.
Analysis deals with the estimation and tests of significance concerning the two parameter α
& β. It does not provide any test as to whether the best functional relationship between X
and Y is indeed linear, because the analysis has been performed under the assumption of
linear relationship between the two.
Steps
149
2) Plot the observed points and draw a graphical representation of the estimated
equation.
3) Test the significance of β.
tb = b ____
√(S2 yx/∑x2)
Regression analysis could be done in several statistical packages. But results from most
common ones are presented below.
MSTATC
There are two option in MSTATC for regression, REGR and MULTIREG. The former is
designed to do a simple linear regression for different groups and compare homogeneity of
regression coefficients. The latter does simple and multiple linear regression.
Example data was taken from soil compaction study available in MSTATC example
directory. Two entries on two compaction methods were used in RCBD with 3 rep. Pod
length, yield, seed/pod, 1000 - seed weight and pods/plant were measured.
For the case of simple linear regression, yield was taken as dependent and pod length as
independent. Output from MSTATC is shown below.
MSTATC gives a lot of output for regression, uncorrected and corrected cross products
variance-covariance, correlation matrix and others. The cross product matrix is used as an
input for further analysis. The regression equation for this data is :
The s.e for the slope is 0.737 which is high since the estimate itself is not very high. Hence,
there is an indication that the estimated slope is not precise. In the ANOVA table, regression
sum of square is not significant at 5% level. Thus t value for b (1.119) indicates that there is
no strong evidence of linear dependence of yield on pod length (p = 0.265).
150
Coefficient of determination is calculated as regression sum of square divided by total sum of
square. For this example R2 = 0.011 (4346.77/413722.59) which is very low, again
indicating poor fit of the model. Here, regression d.f is one because two parameters are to be
estimated and d.f is 2 - 1 = 1.
In the residual table three important results are given: predicted, residual and standardized
residuals.
The predicted column is obtained by substituting pod length and calculating respective yield.
For example, predicted value for case 1 could be calculated as follows:
Residual is defined as the difference between observed value of yield and predicted for a
value of pod length. For example, residual for case 1 could be calculated as:
Large residuals are signals of problem. It simply means that the regression model fitted do
not predict that particular value very well. Hence, there must be a problem with observation
of that particular value in the data set. The value could be outlier or mistyped or could even
be a true value. Therefore, it is good to go back to the data and examine observation with
large residual. For this data, for example, observation no 24,44,76,84, and 94 have large
residuals and need to be re-examined.
151
Data file : COMPACT
Title : SOIL COMPACTION
Function : MULTIREG
Data case no. 1 to 120
10-POD LENGTH
YIELD
4 5
4 6.38679e+003
5 5.26896e+003 4.13723e+005
4 5
4 5.36705e+001
5 4.42770e+001 3.47666e+003
Correlation Matrix
---------------------
4 5
4 1.000
5 0.103 1.000
Intercept = 182.784280
152
RESIDUAL TABLE (sample)
153
SAS
The following SAS program (excluding data step) produces output that follows.
•
•
•
Proc Reg;
Model yield = podlegth/ ss1 p r clm;
Run;
Note that in the model statement commands after the /sign denote an option. For example,
"p" tells SAS to produce predicted values and residuals. If we need to fit regression without
intercept, we use “noint” command after the backslash.
SAS gives ANOVA table, R2, CV and parameter estimate just like MSTATC. If all options
are used, it also give output ranging from ANOVA to model diagnostic tools.
Model: MODEL1
Dependent Variable: YLD
Analysis of Variance
Sum of Mean
Source DF Squares Square F Value Prob>F
Parameter Estimates
Other packages
Regression can be done in several other packages, SPSS, MINTAB, GENSTAT, INSTAT,
etc.
154
11.2 Simple Linear Correlation Analysis
The simple linear correlation analysis deals with the estimation and test of significance of the
simple linear correlation coefficient, r, which is the measure of the degree of linear
association between two variables x and y. The value of r lies within the range of -1 and +1,
with extreme values indicating perfect linear association and mid value of zero indicating no
linear association. The intermediate value of r indicates the portion of variation in one
variable that can be accounted for by the linear function of the other variable. Negative and
positive signs indicates the type of association, i.e, when one variable increases the other
decreases or when one increases the other will increase to or vice versa. Note that the value r
= 0 indicates the absence of linear relationship only. It doesn't show the absence of
relationship, since we are concerned with linear relationship only.
Analysis
Steps.
1) Computer r = ∑xy__
√[(∑x2)(∑y2)]
2) Test the significance of the simple linear correlation by comparing the
computed r value with the tabular r value. r is significant at α level of
significance if the absolute value of the computed r value is greater than the
corresponding tabular r value at the α level of significance.
Correlation analysis is one of the simplest models that can be fitted in a range of statistical
and other packages like Excell. It may be obtained directly when correlation analysis is
submitted or indirectly when other models like regression is fitted. In the example of
MSTATC, for example, correlation between yield and pod length is 0.1 which is not
significant at 5% probability level.
Y1 = a1 + b1x
Y 2 = a 2 + b2 x
155
S2p = (n1-2)S2yx(1) + (n2 -2)S2yx(2)
n1 +n2 - 4
t= b1 - b2
√[Sp2 (1/∑x12 + 1/∑x22)]
Step 4) Compare the computed t with tabular t values with (n1 + n2 - 4) d.f
The above example, regression of yield on pod length, was re-analyzed using compaction as a
grouping variable in MSTATC (REGRE option). In the output below, a lot of information is
given. Mean values, variances and covariance of X and Y is given in a table. Test statistics
and results for the slope (b) for both groups is also given. Slope for compaction 1 is 2.37,
which is significant at 3%, and for compaction 2 is - 0.44, not significant at 5%. The
ANOVA tables test whether regression coefficients are homogenous. Here difference
between levels, i.e intercept is not significant but difference between angles, i.e. slope, is
nearly significant at 5.7% which shows that there is evidence that the slope is heterogeneous.
Therefore, it is possible to conclude that separate regression lines are appropriate for each
group. Hence the equation for each group is given as:
Function : REGR
Data case no. 1 to 120
REGRESSION
X-variable 4 10-POD LENGTH
Y-variable 5 YIELD
Group variables 2
156
From To DF X-BAR Y-BAR VAR.x VAR.y COVAR
----------------------------------------------------------------------------------------
1 60 58 87.43 260.85 49.74 3698.37 118.02
61 120 58 87.98 249.43 58.36 3247.61 -25.52
---------------------------------------------------------------------------------------
Total 118 87.71 255.14 53.67 3476.66 44.28
Within Gr 117 54.05 3472.99 46.25
Between Gr 0 9.08 3910.21 -188.38
From To DF r a b s.b t P%
------------------------------------------------------------------------------------
1 60 58 0.2752 53.4071 2.3726 1.0885 2.1796 0.033
61 120 58 -0.0586 287.9072 -0.4373 0.9779 -0.4472
---------------------------------------------------------------------------------------
Total 118 0.1025 182.7843 0.8250 0.7370 1.1193 0.265
Within 117 0.1067 0.8557 0.7368 1.1613 0.248
Between 0 -1.0000 -20.7576 0.0000 0.0000
When all independent variables are assumed to affect the dependent variable in a linear
fashion and independently of one another, the procedure is called multiple linear regression
analysis. The equation for K independent variables is
157
The data has n(k+1) observation and the layout is
Case no Y X1 X2 ... Xk
1 Y1 X11 X21 ... Xk1
2 Y2 X12 X22 ... Xk2
3 Y3 X13 X23 ... Xk3
. . . . ... .
. . . . ... .
n Yn X1n X2n ... Xkn
Y = α + β1x1 + β2x2
2) Compute
• Residual SS as:
R2 = SSR
∑Y2
158
3) Test the signification of R2
• Compute the F value as:
F= SSR/2
SSE/(n-2-1)
Compare the computed F value with the tabular F value with (k,n-k-1), ie (2,n-3)d.f
The significance of the linear regression implies that some portion of the variability in Y is
indeed explained by the linear function of the independent variables. The size of R2 value
provides information on the size of the portion. Obviously, the larger R2 value is, the more
important the regression equation is in characterizing Y. However, this is not always the case
particularly in a situation where there is multicollinearity between the independent variables.
MSTATC output for multiple regression is given below. From table of parameter estimate it
is clear that seed per pod, seed weight pod per plant significantly contribute to variability in
yield, whereas pod length do not. Normally, if s.e of a slope is greater than its coefficient
estimate then that variable will be dropped even if it is significant. This is because since s.e
is a measure of precision, coefficient with a large s.e will have less precision in estimating
population coefficient, hence may not be included in the model. Therefore, watch out for s.e
in multiple regression.
R2 = 0.89
Function : MULTIREG
Data case no. 1 to 120
10-POD LENGTH
SEED PER POD
1000-SEED WEIGHT
PODS PER PLANT
YIELD
159
Uncorrected
Minimum Maximum Sum Mean Sum of Squares
-----------------------------------------------------------------------------------
4 64.00 106.00 10525.00 87.708 929517. 00
6 3.00 7.00 703.00 5.858 4173. 00
7 148.00 294.00 22659.00 188.825 4319683.00
8 14.00 39.00 2808.00 23.400 69640.00
5 169.00 446.00 30617.00 255.142 8225395.00
------------------------------------------------------------------------------------
Intercept = -423.404368
Variable Selection
So far we only entered the four independent variables and estimated their individual
contributions. We do not know which combination would give best regression model. This
is necessary because some of the independent variables might be highly correlated (multi-
collinearity) and given different results when some are out or in the model. Logically, if two
variables are correlated, then they contribute the same thing to dependent variable, and using
both of them in the model might just be waste of resource and time.
160
that could manage to be included will be published for users. Backward selection is the
reverse procedure. It starts with full model (model that contains all variables) and drop
variable that can't satisfy criteria for staying, until all variables are compared against the
standard set. Stepwise regression, as the name indicates, is a procedure that uses both
forward and backward approaches simultaneously.
Selection methods are not available in MSTATC. Here, SAS was used to run stepwise
selection procedure using same data set. The model statement in the SAS program for simple
regression analysis section may be modified to handle multiple regression.
Proc Reg;
Model yield = podlength podplant seedw /selection = stepwise SLE = 0.05;
Run;
In the first step, SAS selected pod/plant to be entered since it satisfy the criteria, i.e
significance level for entry (SLE) is 0.05. At the second step seed-pod was entered and
pod/plant could not be dropped, etc. The selection procedure checks at each step variables
which satisfy the criteria and that contribute better than all others in the pipeline for entry in
to the model. After a variable is entered it again looks for a variable among those already in
the model, that is short of criteria due to entry of another variable in the model, for exclusion.
Therefore, at each step double-checking occurs, i.e a new variable is entered into the model
and its effect on those already in the model is evaluated. This takes care of multicollineary.
If two variables are highly correlated and if they also highly contribute to dependent variable
obviously both of them can't remain in the model, because if one of them is added to the
model when the other one is already in the model the contribution of the latter given the
former would be so small that it will not make sense at all.
Let's explain the problem of variable selection with a simple example. Assume that the
Biodiversity institute is collecting landraces that best contribute to biodiversity collection of
the nation. It is therefore expected that those landraces collected are unique and could be
considered as an independent wealth of the nation. How do they select landraces then?
Suppose a botanist finds a new landrace and wanted to include it in to the collection. First he
should make sure that this new landrace is not of the same type with any one landrace already
on the shelf. For this he must have mechanisms which we called criteria. Otherwise the
institute may fill all their available space with two or more landraces of same type which do
not uniquely contribute to the collection. Variable selection in regression analysis works
exactly in the same way. We do not require two variables of same nature to be in the model
since both of them will not uniquely contribute to dependent variable.
161
SAS 11:42 Wednesday, September 6, 1989
4
162
Parameter Standard Type II
Variable Estimate Error Sum of Squares F Prob>F
The first technique is linearization, either through transformation or through creation of new
variables. This has two advantages:-
163
transformation or variable creation.
After transformation apply simple linear regression technique. After estimating the
regression coefficients, use them to derive an appropriate estimate of original regression
parameters based on the specific transformation used.
Y X
100 8
216 8.5
193 9.2
94 6.0
208 5.8
306 10.3
177 9.5
290 10.1
256 12.0
285 8.8
Y = αeβx
164
to this data set.
1n (Y) = ln(α) + βX
Which means that Y has to be transformed using the natural logarithm function as follows.
Fitting the data using linear regression (that is using ln(y) as dependent variable and X as an
independent ) the parameters are estimated as follows:
By taking anti - logarithm of 1n(α), α can be estimated to be 31.8. Therefore the curve in its
original scale is given as
Y = 31.8e 0.19x
Such an equation can be linearized by creating K new variables : Z1, Z2, ... Zk, to form
multiple linear equation of the form:
With the linearized form resulting from the creation of new variables, the procedure for
multiple linear regression and correlation analysis can be directly applied..
The most common use of polynomial form is in response curve studies. For this quadratic
polynomial of the form
Y = α + β 1 x + β 2 X2
is often used to locate the optimum response of treatments. A fictious fertilizer trail data
(shown below) was used as an example, where 3 levels of DAP applied and Y yield
harvested. The following data was generated for yield (y) and fertilizer level (x). The
objective is to look for level of fertilizer that gives optimum yield by assuming a quadratic
relationship between yield and fertilizer.
165
Y X X2
8.5 1.2 1.44
9.3 1.2 1.44
8.6 1.2 1.44
It is good to observe pattern of the data using graph to see whether the quadratic form is
valid. Here, yield increased until 2.4 application then steadily declined, which shows a near
quadratic relation. A multiple linear regression of Y on X (X1) and X2 (X2) has been fitted to
estimate b1 and b2. The result was
The next step is to find level of fertilizer that gives optimum yield. This is done using
differentiation principle of calculus
The method is to differentiate Y with respect to X and set the result to Zero. If ∂/∂x is
positive, then the result obtained is a maxima.
Therefore, we may conclude that the maximum yield is obtained for a fertilizer level of 2.45
For functions that can’t be linearized such as population curves and lactation curves for cows,
there is a procedure called NLIN in SAS. One can handle all non-Linear models as long as
the functional form in known. For example, for population growth model y= Aexp[B(x-
1960)] the following program helps estimate the required parameters.
Proc nlin;
Parms A = 3.9
B = 0.022;
Model pop = A* exp[B*( year-1790)];
Run;
166
The non-linear regression often require initial values for parameters to be estimated, because
there could be more than one parameter to be estimated but only one equation is available and
it is not possible to estimate both simultaneously therefore, the package first substitute initial
values for one parameter and then estimate the other. It continues with this turn by turn and
refine the estimate until the system converges..
In the above example, A and B are the ones to be estimated. Hence a range of initial values
has been supplied for SAS to use it. So, we require only “ pop” as a dependent variable in the
model.
For this example, SAS required 8 iterations to have the estimates converged. Here we present
the first two and last two iteration result.
Iter A B SS
0 3.9 0.022 13963.7
1 5.55 0.9196 8577.3
. : : :
. : : :
. : : :
. : : :
. : : :
7 13.998 0.0147 1861.6
8 13.9979 0.0147 1861.6
Here, the iteration for estimation of A converged at the 8th round, because the difference
between 13.9979 is so small that even if 9th iteration has to continue no further refinement is
expected in the value of A .
The NLIN procedure have several options for different results and could be specified.
However, note that non-linear regression is difficult than linear ones to specify and estimate
parameters. Therefore, some complicated models may still be difficult to fit in SAS and you
must watch your result carefully before reporting.
There are four possibilities for the model. They are 1. Separate lines for each levels of the
factor, that is each lines differ from each other both in slope and intercept, 2. Parallel lines
for each levels, that is the lines have same slope but have different intercept so that the rate of
change remain constant, 3. Concurrent line, that is lines which have different slope but have
same intercept. 4. Coincident line, or single line through all data point. The models may be
summarized as follows in the order they appeared above.
167
Model Parameters Residual SS d.f
1 E(yi) = αi + βi Xij α1,α2... αk,β1,β2,...,βk RSS1 n-2k
2 E(yij) = αi + β xij α1,α2, ...,αk,β RSS2 n-k-1
3 E(Yij) = α + βiXij α,β1,β2,..., βk RSS3 n-k-1
4 E(yij) = α + βiXij α,β RSS4
K
Note i = ith levels of factor; j = 1, ... ni , n =∑ ni
i=1
To analyze data using this model the data structure in computer must be clearly presented.
Here we require additional variates to the explanatory variable X which are called dummy
variables, represented by "1" and "0" only. The general data entry structure may be given as
shown below ( in matrix form). It is composed of four parts, the dependent, Y, the K-
dummy variables, the independent X, and the interaction of Dummy by X (D.X).
Two ANOVA tables are possible based on whether parallel or concurrent lines are important.
ANOVA 1
Source df SS
Single line 1 Syy - RSS4
Deviation of paralel K-1 RSS4 - RSS2
from single
Deviation of separate K-1 RSS2 - RSS1
from parallel
Residual n-2k RSS1
Total n-1 Syy
168
ANOVA 2
Source df SS
Single line 1 Syy - RSS4
Deviation of concurrent from single K-1 RSS4 - RSS3
Deviation of separate from concurrent K-1 RSS3 - RSS1
Residual n-2k RSS1
Total n-1 Syy
It was suspected that grain yield of potato (in kg) per plot linearly depend on number of
tubers in a plot. This is because the potato research team believes that tuber size for varieties
X and Y do not differ significantly among themselves. However, it was not known wether
the pattern of linearity is the same for both varieties. Therefore, it was proposed to compute
comparison of regression.
The next step is how to enter the data to the computer. We already have the response and the
explanatory variables, what we additionally need is the dummy variables for both varieties as
the variable "variety" in the dataset above will not help us any more. To generate explanatory
variables (tuber number) for each type of variety we have to multiply dummy by tuber
number (D.X) to get the interaction.
169
Variety 1 Variety 2
Yield (D1) (D2) D1 .X D2. X X
90.5 1 0 6 0 6
88.3 1 0 6 0 6
72.4 1 0 4 0 4
103.6 1 0 12 0 12
99.9 1 0 13 0 13
54.5 1 0 3 0 3
33.6 1 0 2 0 2
48.4 1 0 3 0 3
112.3 0 1 0 15 15
106.0 0 1 0 10 10
93.6 0 1 0 7 7
26.4 0 1 0 2 2
58.9 0 1 0 3 3
77.8 0 1 0 5 5
108.7 0 1 0 9 9
100.5 0 1 0 9 9
The analysis could be done in any statistical package as long as we know which variables we
should fit for the different line types listed above. For example, the following commands in
SAS can analyze the data as required. Two possibilities are available for fitting this model.
Alternative one:
•
•
•
Proc reg data = compreg.reg;
model yield = D1 D2 D1.x D2.x /*for separate line
model yield = D1 D2 X ; /*for parallel line*/
model yield = D1.X D2.X X; /*for concurrent line*/
model yield = X; /*single line*/
In this case D2 will be set to zero for the first two models since it is a linear combination of
D1. The model for separate line will be :
Y = β0 + D1 + D1x + D2x
Hence, for variety 1 : Y = (β0 + D1)+ D1x (note that the intercept is now β0+D1)
170
Y = β0 + D1 + βx
proc Reg;
model yield = D2 D2.x x;
model yield = D2 x ;
model yield = D2.x x ;
model yield = x
Y = β0 + D2 +D2x + B2x
For parallel
Y = β0 + D2 + βx
variety 1: β0 + βx
variety 2: (β0 + D2) + βx
SAS gives ANOVA table for each model fitted and parameter estimate. For example,
ANOVA table for separate analysis (for the first alternative) is as follows.
Model: MODEL1
Dependent Variable: YLD
Analysis of Variance
Sum of Mean
Source DF Squares Square F Value Prob>F
The following parameters have been set to 0, since the variables are a
171
linear combination of other variables as shown.
Let's now construct the two ANOVA tables by subtracting the error sum of square of one
model from another to be able to get extra sum of square.
ANOVA 1
Source df SS MS F
Single line 1 8755 8755 40.3**
Deviation of parallel 1 2706.25 - 2654.03 = 52.22 52.22 0.24
from single
Deviation of separate from 1 2654.03 - 2610.3 = 43.73 43.73 0.20
parallel
Residual 12 2610.3 217.5
ANOVA 2
Source df SS MS F
Single 1 8755 8755 40.3**
Deviation of concurrent from 1 2706.25 - 2614.52 = 91.73 91.73 0.42
single
Deviation of separate from 1 2614.52 - 2610.3 = 4.22 4.22 0.019
Concurrent
Residual 12 = 2610.3 217.0
Once the deviations are obtained the tests are conducted in the usual way, that is by
comparing the mean square of a given model with the error mean square of the full model (a
model that contains largest number of parameters). Consequently, none of the parallel,
separate and concurrent lines are appropriate. Note that the question here is which regression
172
line best represent the pattern in the data set. Now, assuming that a linear relationship
between yield and number of tubers, we are testing additional contribution of parallel,
separate or concurrent as compared to single line through both varieties. Hence, additional
contribution of say parallel lines is a deviation of error SS for single from that of parallel.
For this data set a single line for both varieties were found to be appropriate. Hence, the final
model chosen is
Same analysis could be done in other statistical packages like MSTATC, SPSS, GENSTAT
and others as long as one follows the data entry and analysis steps described earlier.
173
12. Exercises
1. Four treatments, A,B,C, & D are to be used in an experiment. You have a resources
of 25 plots in a rectangular area of this nature.
A
B
C
D
D
Design an experiment!
4. We have 2 varieties, 2N levels and 2 seed rates. Wanted to apply N and seed rate to the main plot
and variety to sub-plot. Produce appropriate split-plot design and sketch ANOVA layout.
7. Consider 81 entries in lattice design. Suppose that Mr. A used 9x9 lattice with 10
replication, Mr B with 5 replication and Mr C with 2 replication. What is the
difference between the three designs.
8. Suppose that you want to conduct an experiment with 3 varieties and 4 fertilizer
level, but could not replicate. Discuss the possible solution for conducting this
experiment. Particularly, is it possible to test effect of main plot with single
replication?
174
13. Further Reading
1. Afifi, A.A and Clark. V. (1984) Computer-Aided Multivariate Analysis.
LifetimeLearning Publication, Velmont California, Chapter 16.
2. B. Jones and M.G Kenward: Design and Analysis of Cross-Over trials.
Chapman and Hall
3. Brian S. Vandell:
Practical Data Analysis for Designed Experiments.
4. Campbell, L.G. and Lafever, H.N (1980).
Effects of locations and years upon relative yields in the soft red winter wheat region.
Crop Sci . 20: 23-28.
5. CC Lindner and C.A Rodger (1997):
Design Theory. CRC press.
6. Cochran W.G (1977) Sampling Techniques 3rd ed, John Wiley & Sons Inc., New
York
7. Cochran, W.G and Cox, G.M (1957). Experimental Designs, 2nd ed,
John Wiley & Sons, New York
8. Cox (1958) Planning of experiments. Wiley, New York.
9. D.K Boniface (1994): Experimental Design and Statistical Method. CRC press, UK
10. Donald W. and J.H. Skillings. (1999) A First Course in the Design of Experiments.
CRC press
11. Federer, W.T (1955) Experimental Design: Theory and Application.
Mac Millan, New York
12. Gomez K.A and Gomez A.A (1984). Statistical Procedures For Agricultural
Research (2nd ed). John Wiley & sons, New York
13. Mead R. (1988). The design of Experiments: Statistical Principles for Practical
Applications. Cambridge University Press
14. Mead R., Curnow R. and A.M Hasted (1999). Statistical methods in Agriculture
and experimental biology (second edition). Oxford University press
15. Trudy A. Watt : Introductory Statistics for Biology Students
16. W.T Federer, Springer, Verlag: Statistical Design and Analysis for Intercropping
Experiment- volume 2 crops.
17. Williams, W.T (1976) pattern Analysis in Agricultural Science.
CSIRO & Elsevire Scientific publishing /co., Melbourne.
18. SAS manual 1999. SAS institute, carry USA.
175