09
DecMONTH START OFFER: Flat 15% Off with Free Self Learning Course | Use Coupon MONTH15
The Analysis of Variance (ANOVA) can be considered as a powerful applied mathematical tool for testing the significance. The test of significance supported t-distribution is an adequate procedure just for testing the importance of the distinction between two sample means. In a situation, once we have three or more samples to think about at a time an alternative procedure is required for testing the hypothesis that each one of the inputs is taken from the same population, i.e. they have the same mean.
Suppose, 5 different fertilizers are used in 4 plots each of which contains wheat and yield of wheat on each of the plots is given. We may be curious about checking whether the effect of those fertilizers on the yields is significantly different or in other words, whether the samples have come from the same normal population. The answer to the present problem is provided by the technique of analysis of variance. Thus, the basic purpose of the analysis of variance is to test the homogeneity of various means.
By ANOVA, the overall variation within the sample data is denoted as the sum of its non-negative components where each of these components is a measure of the variation due to some specific independent factors(causes) separately and then comparing these estimates due to assignable factors(causes) with the estimate due to chance factor(cause), the latter being known as experimental error or simply error.
The steps to design an experiment are:
Experiments have shown that proper consideration of the statistical analysis before the experiment is conducted, forces the experiments to plan more carefully the design of experiments. However, the certainty of the conclusion so drawn regarding the acceptance or rejection of the null hypothesis is given only in terms of probability. Accordingly, the design of the experiment may be defined as "the logical construction of the experiment in which the degree of uncertainty with which the inference is drawn may be well defined".
The basic principles of the design of an experiment are:
Replication can be explained as “the execution of an experiment more than once”. An experiment helps replication to average out the influence of the prospect factors on totally different experimental units. Thus, the repetition of treatment results in a more reliable estimate than is possible with a single observation. The following are the chief advantages of replication:
At the first instance, replication serves to reduce experimental error and thus enables us to obtain a more precise estimate of the treatment effects. From statistical theory, we already know that standard error (S.E.) of the mean of the sample of size n is σ / √n where σ is the standard deviation (per unit) of the population. Thus, if any treatment is repeated r times, then the S.E. of its mean impact is σ / √r, where σ2 is the variance of the individual plot is calculated from the error variance. Consequently, replication has an important but limited role in increasing the efficiency of the design.
The most important purpose of replication is to provide an estimate of the experimental error without which we cannot
As we already discussed, by replication the experimenter tries to average out as far as possible the effects due to uncontrolled effects. This brings to him the question of allocation of treatments to its worth. In the absence of the prior knowledge of the variability of the experimental material, this objective is achieved through “randomization”, a process of assigning the treatments to various experimental units in a purely chance manner. Randomization mainly focuses on:
The validity of the statistical tests of significance e.g. the t-test for testing the importance of the distinction between two means or the ‘Analysis of Variance’, F-test for testing the homogeneity of several means, based on the fact that the applied mathematics into consideration obeys some statistical distribution. Randomization provides a logical basis for that and makes it possible to draw rigorous inductive inferences by the use of statistical theories based on probability theory.
The purpose of randomness is to assure that the sources of variation, not controlled in the experiment, operate randomly so that the average effect on any group of units is zero.
It should be kept in mind that performing randomization only without performing replication is not enough. It is only when randomization of treatments to various units is accompanied by a sufficient number of replications then we are in a position to apply the test of significance.
If the experimental material, say field for agriculture experimentation, is heterogeneous and different treatments are allocated in various units(plots) at random over the entire field, the soil heterogeneity will also enter the uncontrolled factors and therefore increase the experimental error. It is mandatory to decrease the experimental error as much as possible without increasing the number of replications or without interfering with the statistical requirement of randomness so that the fewer differences between treatments can also be detected as important. The experimental error can be reduced by making use of the fact that neighbouring areas in a field are relatively more homogenous than those widely spread. To separate the soil fertility effects from the experimental error, the whole experimental area is divided into homogenous groups(blocks) row-wise or column-wise (one-way elimination of fertility gradient, c.f. Randomized Block design) or both (elimination of fertility gradient into two perpendicular directions c.f. Latin square design, according to the fertility gradient of the soil such that the variation within each block is minimum and between the blocks is maximum. The treatments are then assigned at random inside every block.
Read: How to Do Data Manipulation of Packages Using R?
The method of reducing the experimental error by dividing the comparatively heterogeneous experimental area(field) into homogenous blocks (due to physical closeness as for as field experiments are concerned) is known as Local control.
The Analysis of Variance (ANOVA) is mainly classified into two categories:
But here we are only discussing one-way classification.
Let us suppose that N observations x_{ij} (i=1,2,…..,k; j=1,2,…., n_{i} ) of a random variable X are classified on some basis into k categories of sizes n_{1}, n_{2},……., nk, (N=i=1kni ) as exhibited below:
Means | Total | ||
X_{11} | X_{12} …….x_{1n1} | x̄_{1} | T_{1} |
X_{21} . . . . | X_{22} …….x_{2n2} . . . . . . . . | x̄_{2} . . . . | T_{2} . . . . |
X_{i1} | X_{i2} …….x_{ini} | x̄_{i} | T_{3} |
X_{k1} | X_{k2} …….x_{knk} | x̄_{k} | T_{4} |
The total variation within the observation xij can be categorized into the following categories:
The first variety of a variable is because of negotiable causes which can be detected and managed by human endeavour and the second variety of variation is because of likelihood causes that are on the far side of the control of the human hand.
The main target of the ANOVA technique is to check whether there is enough difference between the class means because of the inherent variability inside the separate categories.
The linear mathematical model used in one-way classification is:
x_{ij }= µ_{i} + i_{j}
= µ + (µi - µ) + £_{ij}
= µ + i + £_{ij }
where, (i=1,2,…..,k ; j=1,2,….,ni ),
x_{ij} is the yield from the j^{th} cow, (j=1,2,…,ni) fed on the ith ration (i=1,2,…,k),
µ is general mean effect given by
µ=i=1kni µi / N,
Read: Data Science Tutorial Guide for Beginner
a_{i} is the effect of the ith ration given by a_{i} = µ_{i} - µ, (i=1,2,..k),
£_{ij} is the error because of chance
Statistical analysis of the Model
Hence,
Total S.S. = S.S.E + S.S.T
Degree of freedom for various S.S.
ST^{2}, the total S.S. which is computed from N quantities of the form (x_{ij} - x̄..)^{2} will carry (N-1) degrees of freedom(d.f.).
Mean Sum of Squares (M.S.S.)
When we divide S.S by d.f then it gives M.S.S.
ANOVA Table for One-way Classification
Sources of Variation |
Sum of Squares |
d.f |
Mean Sum of Squares Read: ARIMA like Time Series Models and Their Autocorrelation |
Treatment (Ration) |
S_{t}^{2} |
k-1 |
s_{t}^{2}= S_{t}^{2}/(k-1) |
Error |
SE^{2} |
N-k |
S_{E}^{2}= S_{E}^{2}/(N-K) |
Total |
ST^{2} |
N-1 |
Conclusion
In this blog, we have discussed the Analysis of Variance (ANOVA). It is a statistical tool for the testing of significance. We have discussed the method for designing an experiment and also the classification of ANOVA. Further, One-way ANOVA is explained in detail, its approach, mathematical model, different sum of squares, degree of freedom and mean sum of squares.
Please leave the query and comments in the comment section.
Read: How to Work with Regression based Models?
A dynamic, highly professional, and a global online training course provider committed to propelling the next generation of technology learners with a whole new way of training experience.
AWS
DevOps
Data Science
Hadoop
Salesforce
QA
Business Analyst
MS SQL Server
Python
Artificial Intelligence
Machine Learning
Tableau
Search Posts
Trending Posts
Top 30 Core Java Interview Questions and Answers for Fresher, Experienced Developer 32.3k
Cloud Computing Interview Questions And Answers 25.2k
Difference Between AngularJs vs. Angular 2 vs. Angular 4 vs. Angular 5 vs. Angular 6 23.2k
What is SFDC? What does SFDC stand for? 21.6k
SSIS Interview Questions & Answers for Fresher, Experienced 18.6k
Related Posts
Receive Latest Materials and Offers on Data Science Course
Interviews