Sei sulla pagina 1di 5

ANALYSIS OF VARIANCE (ANOVA)  F TEST USING DEVIATION METHOD:

 RONALD AYLMER FISHER (R.A. FISHER) -He developed Analysis of Variance To solved for F, the means squares (MS) should be computed.
(ANOVA) in 1923. For the mean square within (MSw ), we use the equation:
s21 +s22 +s23 … +s2k
- The F-test used in ANOVA is named MSw= Where:
k s 2 = group variance
after him. k= number of sample or groups
 ANALYSIS OF VARIANCE (ANOVA) - It is used to test hypothesis about the For the mean square between (MSb ), we use the equation:
differences between two or more means. ∑(x̅−x
̅̅̅̅)
G
2
Msb = m Where:
-The key statistics in ANOVA is the F-test of k−1
m= number of observations in each sample/group
difference of group means. xG 2 = summation of the squared deviation between the
∑(x̅ − ̅̅̅)
grand mean and the mean of each group/sample
k= number of samples/ groups
 ONE WAY ANOVA - Tests differences in SINGLE DEPENDENT VARIABLE among
two or more groups formed by the categories of a SINGLE CATEGORICAL To determine if the computed F value is significant or not, the critical should
INDEPENDENT VARIABLE. be known. To find the critical value, compute for the degrees of freedom first. To
-Also known as Univariate ANOVA, Simple ANOVA, Single compute for the degrees of freedom use:
classification ANOVA, or One-factor ANOVA. Dfb = k − 1 Dfw = N − k
-It tests whether the MEAN of groups formed by the 1. S𝑆𝑇 = (𝑡𝑜𝑡𝑎𝑙 𝑠𝑢𝑚 𝑜𝑓 𝑠𝑞𝑢𝑎𝑟𝑒𝑠)
categories of the independent variable seem similar. ∑(𝑥𝑡 )2 Where:
-If the groups seem different, then it concludes that the sst = 𝑥 2 𝑡 − ∑XT = summation of the total of all the groups
𝑁 ∑XT2 =
Independent variable has an effect on the dependent variable. summation of the squared scores of all the groups
2. S𝑆𝑏 = (𝑏𝑒𝑡𝑤𝑒𝑒𝑛 𝑠𝑢𝑚 𝑜𝑓 𝑠𝑞𝑢𝑎𝑟𝑒𝑠) N = total sample size
n= sample size of each group
 F – TEST - is the ratio of variance within and variance between of the groups. (∑ x1 )2 (∑ x2 )2 (∑ x3 )2 (∑ xk ) 2
ssb = [ + + + ]
- the ratio of the variable estimates. n1 n2 n3 nk
𝑀𝑆 2
F=𝑀𝑆 𝑏 Where : (∑(xT )
𝑤
M𝑆𝑏 = mean square between −
M𝑆𝑤 = mean square within
N
To solve for F, we have to solve for the sum of squares(SS) using the :
a. Deviation method 3. SSw = (within sum of squares)
b. Raw score method SSw = SST − SSb

Solve for the mean squares (MS)


SSb SSW
for MSb = dfb
for MSW = dfw

For degrees of freedom (df) so,

Dfb = k − 1 Dfw = N − k
HYPOTHESIS TESTING  Application of Z-Test - used if sample size is n≥30
Steps and Procedures to be followed in using Z-test
 HYPOTHESIS - is concerned with the decision making rules for choosing 1. Formulate null and alternative hypothesis
alternatives while controlling and minimizing the risk of wrong decisions. 2. Set the level of significance and decide whether a one tailed test
 NULL HYPOTHESIS (H0) - It is a preliminary assumption about the parameter of a or two-tailed test
population/s that is being investigated. It is a claim that a population 3. Decide the test statistic to be used
characteristic that is initially assumed to be true. Null hypothesis is tested against 4. Compute for the value of the test statistic using sample data
another hypothesis known as the alternative hypothesis 5. Make a decision
 ALTERNATIVE HYPOTHESIS (Ha) - Is a researcher's formulation of the hypothesis a) If Z > Zt then reject the null hypothesis
or assertion which he/she believes to be true about the population parameters. b) If Z< Zt then accept the H0
Ha is an assertion contradicting the null hypothesis. 6. State the conclusion
 HYPOTHESIS TESTING - It is used to determine which hypothesis is more
acceptable as true or which hypothesis is more likely to be false.
 METHODS USED TO TEST HYPOTHESIS-  APPLICATION OF T- TEST - Used if sample size is small (n< 30)
o Traditional or Classical Steps and Procedures to be followed in using T-test
o P-Value 1. Formulate null and alternative hypothesis
o Confidence Interval method 2. Set the level of significance and determine the degree of
freedom; find the tabular value or critical value df = n-1 (for one
sample mean test) df = n1+n2 -2 (for 2 groups of samples)
3. Decide the test statistic to be used
4. Compute for the value of the Tcomputed
5. Decide whether to accept or reject the null hypothesis
6. State the conclusion
SIMPLE REGRESSION AND CORRELATION ANALYSIS
 REGRESSION ANALYSIS - It is useful in determining the possible from of  REGRESSION LINE- Is a line drawn through the points on a scatter plot to
relationship between variables to predict or estimate the value of another summarize the relationship between the variables being studied.
variable.  NEGATIVE OR INVERSE RELATIONSHIP BETWEEN THE VARIABLES - When
 SCATTER DIAGRAM- Is a graph in which the values of two variables are plotted the line slopes down (from top left to bottom right).
along two axes, the patter of the resulting points revealing any correlation present.  POSITIVE OR DIRECT RELATIONSHIP - When it slopes up (from bottom
 CORRELATION ANALYSIS- Refers to the method of measuring the strength of left to top right.
association (correlation) among variables.  LINEAR REGRESSION - Is used to make predictions about a single value.
 POSITIVE CORRELATION COEFFICIENT - (0<r<1) When one variable measures, as  SIMPLE LINEAR REGRESSION - Aims to find linear relationship between a
the other measure, the correlation is positive. response variable and a possible predictor variable by the least square
 NEGATIVE CORRELATION COEFFICIENT - (-1<1<0) When one decreases as the method.
other increases, it is negative
 NO CORRELATION - (r=0.00) Complete absence of correlation is represented by 0  REGRESSION EQUATION- Is a mathematical equation that is used to
0.0 -no correlation/negligible correlation
predict the values of one dependent variable from known values of one
±1.00 -perfect correlation
or more independent variables.
± 0.01- ± 0.25 -very low correlation  Dependent Variable- is a variable being predicted or
± 0.26- ± 0.50 -moderately low correlation explained.
± 0.51- ± 0.75 -high correlation
 Independent Variable is the variable that is used to predict
± 0.76- ± 0.99 -very high correlation
or explain the dependent variables.
 PEARSON’S PRODUCT INTERVAL VARIABLES - It is the measure of the of the linear ∑ 𝑥𝑦 –𝑛𝑥̅ 𝑦̅
𝑦̂ = a + bx 𝑎 = 𝑦̅ – b𝑥̅ b=∑
association between two variables that are measured on interval of the ratio 𝑥 2 −𝑛 ∑ 𝑥̅ 2
𝑁 ∑ 𝑥𝑦−(∑ 𝑥)(∑ 𝑦)
scales.  REGRESSION ANALYSIS- Used to develop a statistical model that can be
√[𝑁 ∑ 𝑥 2 −(∑ 𝑥)2 ][𝑁 ∑ 𝑦 2 −(∑ 𝑦)2 ]
used to predict the values of one variable based on the values of another
 SPEARMAN RANK ORDER CORRELATION COEFFICIENT- Is used to calculate the variable.
correlation of ordinal data which are classified according to order or rank.
 COEFFICIENT OF DETERMINATION(𝒙𝟐 ) - is a key output
6(∑ 𝐷 2 )
𝑟𝑠 =1-𝑁(𝑁2 −1) of regression analysis.

Coefficient of Determination= (𝐂𝐨𝐫𝐫𝐞𝐥𝐚𝐭𝐢𝐨𝐧 𝐂𝐨𝐞𝐟𝐟𝐢𝐜𝐢𝐞𝐧𝐭)𝟐


𝒓𝟐 =(𝒓𝒙𝒚 )𝟐
THE CHI-SQUARE TEST Table of Observed Frequencies
 Karl Pearson (1857-1936) -first used the chi-square distribution as a 𝑎1 𝑎2 𝑎3 G

goodness-of-fit test for data.


𝑏1 𝑏2 𝑏3 H D,E & F – represent the sub totals of the observed
 THE CHI-SQUARE TEST- It is used to analyzed discrete data. frequencies in every column.
 THE CHI-SQUARE GOODNESS OF FIT TEST - can be used to see G,H & I – represent the sub totals of the observed
𝑐1 𝑐2 𝑐3 I
whether a frequency distribution fits a specific pattern. It involves frequencies in every rows.
T – represent the grand
only single variable with two or more categories. D E F T total.
 Formula for computing the test value:
𝑘 where:
(𝑂𝑖 − 𝐸𝑖 )2
𝑥2 = ∑ O = observed frequencies
𝐸𝑖 E = expected frequencies
Table of Expected Frequencies
𝑖=1
k = number of categories
𝐷𝐺 𝐸𝐺 𝐹𝐺
with df = k – 1 𝑇 𝑇 𝑇
D,E & F – represent the sub totals of the observed
 THE CHI-SQUARE GOODNESS OF FIT TEST
1. State the hypothesis and identify the claim. 𝐷𝐻 𝐸𝐻 𝐹𝐻 frequencies in every column.
𝑇 𝑇 𝑇
2. Find the critical value. The test is always right-tailed G,H & I – represent the sub totals of the observed
3. Compute the test value. Find the sum of the values. 𝐷𝐼 𝐸𝐼 𝐹𝐼 frequencies in every rows.
(𝑂𝑖 − 𝐸𝑖 )2 𝑇 𝑇 𝑇 T – represent the grand total.
𝐸𝑖
 THE CHI-SQUARE GOODNESS OF INDEPENDENCE
4. Make the decision.  Formula for computing the test value:
where: with df = (r-1)(c-1)
5. Summarize the results. 𝑘 2
(𝑂𝑖 − 𝐸𝑖 ) O = observed frequencies
𝑥2 = ∑ where: r = number of rows
 THE CHI-SQUARE TEST OF INDEPENDENCE - The Chi-square Test of 𝐸𝑖 E = expected frequencies c = number of columns
𝑖=1 k = number of categories
independence is used on analyzing two variables with multiple
categories. When there are already more than one variable involved in a  THE CHI-SQUARE GOODNESS OF FIT TEST
study, the chi-square goodness of fit can no longer be applied. 1. State the hypothesis and identify the claim.
2. Find the critical value. The test is always right-tailed
3. Compute the test value. Find the sum of the values.
(𝑂𝑖 − 𝐸𝑖 )2
𝐸𝑖
4. Make the decision.

5. Summarize the results.

Potrebbero piacerti anche