Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Student ID : 8180021
September 2019
Multiple Liner Regression
Contents
1 General Overview of The Project Problem ................................... 3
Model 1: .................................................................................................................................... 8
Model 2: ...................................................................................................................................15
Model 3: .................................................................................................................................. 22
5 Appendices ............................................................................. 48
2
Multiple Liner Regression
The Experiment that used in this project is a Number of Fish In Water Stream [Y]
calculating multiple factors:
Moreover, the maximum depth of the stream will be measured along the cross-sectional
width.
1- Dissolved Oxygen:
2-4 mg/L only a few fish and aquatic insects can survive.
4-7 mg/L good for many aquatic animals, low for cold water fish
2- Nitrate No3:
Nitrate is measured in mg/L. Natural levels of nitrate are usually less than 1 mg/L.
Concentrations over 10 mg/L will have an effect on the freshwater aquatic environment. 10
3
Multiple Liner Regression
mg/L is also the maximum concentration allowed in human drinking water by the U.S.
Public Health Service. For a sensitive fish such as salmon the recommended concentration
is 0.06 mg/L
3- Temperature:
In hot weather, the first thing to do is to increase water movement. The warmer water
is, the less oxygen will be dissolved in it, but at higher temperatures, your fish’s
metabolism will be higher, increasing their need for oxygen.
5-Depth of Stream:
Stream Channel depth vary along the stream cross section, the experiment calculate the
maximum depth in the stream cross-section, the fish does not prefer shallow stream water
to be there
4
Multiple Liner Regression
6-Sulfate: Sulfate has effects on fish at high concentration only, when reach
hundred mg/l.
All the Above independent variable will model in order to find their relation with the
dependent variable which is the Number of Fish the stream [Y]
All Variable is continuous variable, in the last section we will convert the independent
variable into binary categorical variable and convert one dependent variable into binary
categorical variable
5
Multiple Liner Regression
6
Multiple Liner Regression
As shown above from descriptive statists table we have a quick review for the data and we
can see that we have 68 Sample Case, we can also see the mean and Std deviation and
compare them together. For Example, [X4] Has mean Value of 5.6, and it Std deviation
is equal to 3.27 this is massive STD comparing to the mean!!
As Shown Above is the Descriptive Statistics for each Variable and Normality Test for each
for them using Kolmogorov-Smirnov and Shapiro-Wilk.
7
Multiple Liner Regression
Model 1:
Model 1 is the first trial model that contains all
variables together that entered in Forced Entry
method and assumed this model have constant
[have intercept with Y Axis], the result is as
following
We can see that all Variable is entered without any exclusion, therefore the R2 for the
model will be for all these variable with [Y]
8
Multiple Liner Regression
This F Test Represent us a statistics test for the overall significance of the regression model, we
always seek to High F Value, and seek a value of Sig < 0.005 to state that the observed
difference is real not due to chance.
In-our Model Sig is equal to 0.000 which is less than 0.005, so that the null hypothesis is
rejected and the observed difference is real not due to chance.
9
Multiple Liner Regression
From Coefficients table we can see B value for each variable and whether it is significant or
not.
We can see from the coefficients table that Constant, [X5] , [X6] have Sig Higher Than
0.05 therefore the null hypothesis accepted [ this coefficients is bad in our model ]
On other hand it seems that [X1] ,[X2],[X3],[X4] have a Sig Value less than 0.05 , therefore
the null hypothesis is rejected and they are well fit in the model .
We Can interpret from coefficients table the VIF Values that help us to identify Collinearity
First of all, we don’t have any variable have VIF > 10, on other hand all Predictors have VIF
around 1, expect The Water Temperature [X6] and Dissolved Oxygen [X2] both have 3.187
and 3.451 respectively. As first indication these to two variables Temperature and dissolved
Oxygen maybe have a multi-collinearity problem, but let us take a look in the coloration
and covariance matrix.
10
Multiple Liner Regression
From The Coefficients Correlations we can that all variable have small values of r between
them expect Water Temperature [X6] and Dissolved Oxygen [X2] they have correlation
0.821 !!!!, this impressive result is correspond with VIF in Page 10, so that Water
Temperature and Dissolved Oxygen have high probability to be multi-collinear variable,
Maybe the next Model [ Model 2] [Backward] will eliminate one of them if one of them does
not add any significance effect on the model. Because in multi-collinearity one of the
variables is enough in model, and the other one is redundant and we can eliminate one of
them from model.
11
Multiple Liner Regression
The condition indices are computed as the square roots of the ratios of the largest
eigenvalue to each successive eigenvalue. Values greater than 15 indicate a possible
problem with collinearity; greater than 30, a serious problem. [ In our Model We have only
one parameter which is temperature is have Condition Index 53.511 > 30.
Eigen Value Problem Is that it does not have cutoff point to judge with it.
Case-wise: There is only two cases in the model which is consider as outliers because these
two cases have Std Residual > 1.96, 2 Cases Over 68 Cases is equal to 2/68 = 2.94%, which
is fine 2.95% < 5%. In My Opinion I don’t like to remove outliers in-case std residuals at 2
Std is less than 5%, Because sometimes removing outliers is an endless process every
time you will remove outliers the model will adjusted and find a new outliers according the
new fit model and so on.
I’m trying to remove these residuals from the model, and I have a three new residuals, so
that my theory is correct about if only 5% of your Std Residual or less lays below 2 Std
deviation then you are Ok!
12
Multiple Liner Regression
From Scatter Plot we can find that the data here is Heteroscedasticity, so that the residual
increase with increase of predicted value, the variance of the residuals is not constant at
each level of predictor variables. No Clear Evidence on linearity or not
From The Test of normality above we can see that all residuals type is normally distributed
because it has Sig > 0.05.
13
Multiple Liner Regression
1-Cooks distance, it should be < 1, if not (>1) then this case in concerning
From our model no case have cooks distance more than 1 , and this is perfect
2-Leverage: average leverage in our model 1 = (K+1)/n, (6+1)/68 = 0.103 cases with 3 times
average leverage 3 time is concerning 3 X 0.103=0.309 .
From our model only case 28 have leverage value more than 0.309 but is cooks distance is
less than 1 so we can consider it as ok !
No DF Betas are more than 1 expect case No.1 but it is slightly near 1 and its cooks distance
is also fine.
3∗(𝑘+1) 3∗(𝑘+1)
4-Covariance Ratios , values greater than 1+ and less than 1- are concerning
𝑛 𝑛
3∗(6+1)
, then 1± , then the range between [0.691 to 1.309 ] is acceptable range , any case out
68
Covariance Ratios only case 1,7,22,28,38,48,57 have values of CVR out of this range
[0.691 to 1.309], but also their cooks distance is also fine less than 1.
5-Mahalanobis distance: for our sample small one (68)<100 , any values above 15 need to
take care of it . only case 12,22 have value higher than 15, but their cooks distance is also
fine
AS a conclusion the model has not have many influential point and there are too little
indicators for them and because all of them does not have cooks distance more than 1
14
Multiple Liner Regression
Model 2:
As we can see the model start with full variable then it removes the water temperature
because it has Sig 0.983 > 0.1, after that it removes Sulfate Concentration because it has Sig
0.921 > 0.1, after than the model stop removing because any variable removed it will have a
Sig < 0.1, so its removal will be significantly change the model!
15
Multiple Liner Regression
As we can see from the model summary the model the 1st model it starts with 0.624 and
removing 2 variable [X6] and [X5] will not affect the model, but it is improving the
difference between R Square and Adjusted R Square Now R square 0.624 and adjusted R
Square 0.601. , Durbin Watson for independent error check is 1.018 > 1 and that is ok
16
Multiple Liner Regression
ANOVA Test Also Shows us how F Improve through model elimination process we always
seek for Higher F Value with Sig <.05 , and in this model Sig < 0.05 , So it is very good
As shown in coefficients table now in model [2] all variables have significance less than
0.05 and that is perfect, now only constant has Sig Higher than 0.05. Moreover No VIF >10
, and All VIF is Around 1 that result is away better than Model [1] .
17
Multiple Liner Regression
The condition indices are computed as the square roots of the ratios of the largest eigenvalue
to each successive eigenvalue. Values greater than 15 indicate a possible problem with
collinearity; greater than 30, a serious problem. Six of these indices are larger than 30,
suggesting a very serious problem with collinearity. [ In our Model We have any only [X4]
with CI > 15, No Value More than 30 .
Same as model [1] the same outliers appear but they only represent less than 5% from all sample
so outliers is fine .
Side Note : there is no outliers more than 3 Std
18
Multiple Liner Regression
From Scatter Plot we can find that the data here is Heteroscedasticity, so that the residual
increase with increase of predicted value, the variance of the residuals is not constant at
each level of predictor variables. No Clear Evidence on linearity or not
We can see from test for Normality that all residuals are normally distributed with Sig >
0.05, and that is perfect
19
Multiple Liner Regression
1-Cooks distance, it should be < 1, if not (>1) then this case in concerning
2-Leverage: average leverage in our model 1 = (K+1)/n, (6+1)/68 = 0.103 cases with 3 times
average leverage 3 time is concerning 3 X 0.103=0.309
3∗(𝑘+1) 3∗(𝑘+1)
4-Covariance Ratios , values greater than 1+ and less than 1- are concerning
𝑛 𝑛
3∗(6+1)
, then 1± , then the range between [0.691 to 1.309 ] is acceptable range , any case out
68
5-Mahalanobis distance: for our sample small one (68)<100 , any values above 15 need to
take care of it .
As a conclusion the model has not have many indicators for influential points and all cases
does not have cooks distance more than 1
20
Multiple Liner Regression
Side Note:
Model [2] : Can be Run through Forward Entry Method , and it will have the same result of
backward in this model , same R Square , same coefficient , same everything ) , but the
procedures that forward run out it is totally different way from Backward .
Here is how the forward for our model will be run out
As we see here the model start without any variables, then it is add the most variable will
give us a significant change in R Square with Sig < 0.05, after that [ addition of Maximum
water Depth], it is found that [nitrate is the most second variable] that if added will improve
the model, and so on …..
In this experiment Forward and backward model give the same result Same R2 and same
regression equation
21
Multiple Liner Regression
Model 3:
Based on the above if there is no area provide stream with water [X1=0] then there is no
stream at all then no fishes !! . Same concept can be applied on water depth if the stream
depth is zero [X3=0] then there is no water then no
fishes!!
22
Multiple Liner Regression
As we discussed in model 2 , the variable with Sig > 0.1 will be excluded , because their
removal from the model will not affect the model significantly
Sulfate and temperature excluded because they have Sig > 0.1
Wow !!!, Removing constant have massive improve in R Square it is jump from 0.624 to 0.914
and the difference between R Square and Adjusted R Square is decreased, on other hand
Durbin Watson here is less than 1 , but it is is too close to 1 , 0.98 approximately 1 .
23
Multiple Liner Regression
ANOVA Test All F Values have Sig < 0.005 That is Very Good, and F is increasing per each
iteration
24
Multiple Liner Regression
Look how the coefficient is Perfect! all Coefficients is Significant, and well fitted to the
model, all coefficients have sig value less than 0.05.
The Problem that appears here that we have two variables with VIF > 10, Dissolved Oxygen
and Maximum Water Depth this may lead to multi collinearity
The Only Problem in modeling through origin that the variable it seems to be multi-
collinear with high person coloration between themselves.
So That we need to keep any eye on high VIF Variable because it has also high correlation [
Keep an eye on Dissolved Oxygen and Maximum water depth] or remove one of them from
the model
25
Multiple Liner Regression
The condition indices are computed as the square roots of the ratios of the largest
eigenvalue to each successive eigenvalue. Values greater than 15 indicate a possible
problem with collinearity; greater than 30, a serious problem. Six of these indices are larger
than 30, suggesting a very serious problem with collinearity. We don’t have any Varible with
CI > 15.
We can see from test for Normality that all residuals are normally distributed with Sig >
0.05, and that is perfect .
26
Multiple Liner Regression
From Scatter Plot we can find that the data here is Heteroscedasticity, so that the residual
increase with increase of predicted value, the variance of the residuals is not constant at
each level of predictor variables.
No Clear Evidence on linearity or not
27
Multiple Liner Regression
1-Cooks distance, it should be < 1, if not (>1) then this case in concerning
2-Leverage: average leverage in our model 1 = (K+1)/n, (6+1)/68 = 0.103 cases with 3 times
average leverage 3 time is concerning 3 X 0.103=0.309
3∗(𝑘+1) 3∗(𝑘+1)
4-Covariance Ratios , values greater than 1+ and less than 1- are concerning
𝑛 𝑛
3∗(6+1)
, then 1± , then the range between [0.691 to 1.309 ] is acceptable range , any case out
68
5-Mahalanobis distance: for our sample small one (68)<100 , any values above 15 need to
take care of it .
AS a conclusion the model has not have many influential point and there are too little
indicators for them and because all of them does not have cooks distance more than 1
28
Multiple Liner Regression
3 Logistics Model:
In Logistics Model we are going to Convert the continuous dependent variable into a Binary
categorical variable by follow this category:
If the number of fish range from [ 0 – 60] then it is a Poor Stream with Aquatic life. [0]
If the number of fishes is more than 60 then it is a rich stream with Aquatic life. [1]
If the Area Drained by Steam is More than [ 7000] Acres, then it is a large area [1]
29
Multiple Liner Regression
30
Multiple Liner Regression
Logistic Model 1 :
By Forced Enter Method , We enter all variable , and make [X1] as Independent categorical
variable and make dummy coding by categorical option in binary regression .
31
Multiple Liner Regression
Classification Table gives us a fast indication how the percent of model correctness in block zero in
our model it is 75% .
-2 L L = 76.478 , We Always Seek for minimum log likelihood , this number should be improve and
go down when model Block 1 .
Exp[B] is the Odd Raito this number should be far from 1, because, this number means how 1 unit
change in this variable will affect the independent variable probability of the outcome that coded
“1” to happen in our case “ Rich Aquatic Life “ Coded as “1” , Value of 3 is very good.
the null hypothesis is that the two coefficients of interest are simultaneously equal to zero. If the
test fails to reject the null hypothesis, this suggests that removing the variables from the model will
not substantially harm the fit of that. model
Then Sig = 000 then Sig < 0.05 we reject the null hypothesis so this is good one
32
Multiple Liner Regression
Block 1:
AS we see each iteration the -2Log Likelihood is decreasing and this is what we want we want the
minimum 2log likelihood from the model
As we see from classification table the correctness percent increase from 75% to 95.6%, and it is
very good percent
As we can see also the model predict a 100% percent of cases in Rich Aquatic life , 51 out of 51
and 82.4 % from poor Aquatic Life 14 out of 17.
From Model Summary we can see that the new -2Log Likelihood is 17.632 which decreased from
the value in Block zero and that is good, Also we can here see both R Square Terms that equivalent
to the R Square in linear Regression Cox Snell = 0.579 and Nageelerke = 0.858 and both of them
depends on -2Log Likelihood. both of them can be consider as good R Square.
33
Multiple Liner Regression
One of the most important test for the model fitness using Chi-Square Distortion the
Hosmer And Lemeshow Test, is testing
The Null Hypothesis is: the observed and expected proportions are the same across all
doses [ this mean well fitted model]
Alternative is: The alternative hypothesis is that the observed and expected proportions are
not the same
Therefore, we would like to accept the null hypothesis, in our model Sig = 0.89 > 0.05 , then
accept the null hypothesis that the model is well fitted
Omnibus Test is also good to know how significance the 2LL reduction is significant or not
for example in block 2LL was equal 76.47 and then in block 1 it was reduced to 17.623 is this
reduction from 76.47 to 17.623 significance or not??? Omnibus answer this question.
the null hypothesis of Omnibus test is : it reduction in 2LL from bassline model is not significant
In our model sig form omnibus test is 0 , which is < 0.05 so that we reject the null hypothesis
and accept that the reduction in 2LL is significant
34
Multiple Liner Regression
Now Let take a look for Variable in Equation first of All we note that [x3] [x5] [x6] [X2] and
constant has Sig > 0.05 and this is bad. On other hand we have [X1] [X4] have the Sig < 0.05.
Another problem here for [X5] and [X3] the EXP(B) [Odds Ratio] is too close from 1, and this
is bad sing this mean that increasing 1 unit in this variable will not change the probabilities
of occurrence for “1” (Rich Aquatic life).
Another Problem Also is The Range of C.I This interval should not have passing throughout
[1] on it only variable that meet this criteria is [X1] [X4] , and other variable is all passing
through 1 and this is also bad sing
We can see from this plot that most value is clustered at edged and this is perfect because
we don’t want value near the 50% probability cutoff point, when value around 50% increase
the error increase in model but forint in our model the value scatter at edge.
35
Multiple Liner Regression
Linearity has been met because all Sig > 0.05 that is amazing.
36
Multiple Liner Regression
1- Cooks Distance >1 , we have 3 cases that have cooks distance more than 1 , Case 5 , Case 12
And Case 17 , we should care about these because these cases are outliers as shown in case
wise and also they are Influential Point .
2- Leverage: average leverage in our model 1 = (K+1)/n, (6+1)/68 = 0.103 cases with 3
37
Multiple Liner Regression
All VIF Are Less Than 10, so this is fine, but maybe we need to take care of Temperature and
dissolved oxygen, because they are already have high correlation in the linear regression
process
Collinearity Statistics
1 (Constant)
38
Multiple Liner Regression
39
Multiple Liner Regression
Block 1:
We can see that the correctness of our model is increasing by each step in forward method
our model final correctness is 95.6% [ Same As Model 1 ] as the forced entry method entry
but there is some variable is eliminated
We can notice from the model summary from block 1, that each step has a lower -2Log
Likelihood value this mean that each step has a better model than the step before and we
can notice that in step 3 2 Log the minimum value that we have possible get.
The step in forward start with empty model then it is adding the most significance variable
in the model after than it is add the second significance variable, the model stops when the
reduction by adding extra variable in in -2Log Likelihood is not significant anymore.
Our model have Cox Snell R – 0.573 and Nagelkerek R = 0.849 this R Squares is calculated
throughout -2 Log Likelihood [ Same As Model 1 ]
40
Multiple Liner Regression
As we see the model start with [X4] , because it is the most significant from all other
variable , then in step 2 the model add [X1] , then the model add [X2] , after that the model
stop because any addition of the other reaming variable will not improve the model
significantly all remaining variable [x3][X5][X6] is all have Sig > 0.05 because of that
there is no addition happens .
From Variables in equation table we can find the following: [X2] [X4] [X1] have sig < 0.05,
this model has better coefficient significance than the Model 1 In page
41
Multiple Liner Regression
35 Good News here! is that EXP(B) CI No Passing 1 on all Variables !! that is amazing and
better than model Number 1 in the page 35
The Null Hypothesis is : the observed and expected proportions are the same across all
doses [ this mean well fitted model ]
In Our Model we have in step 3 Sig = 0.763, so that we will accept the null Hypothesis and
the model is well fitted
the null hypothesis of Omnibus test is: it reduction in 2LL from bassline model is not
significant
In our model sig form omnibus test is 0, which is < 0.05 so that we reject the null hypothesis
and accept that the reduction in 2LL is significant
42
Multiple Liner Regression
We can see throughout these chart how our model is well scatter at the edge this mean that the
model is very good one, and we don’t have any values around 50% the cutoff point, if the model has
many values around 50% this mean the model is a bad one
43
Multiple Liner Regression
Influential cases
1- Cooks Distance There is only one case with cooks distance > 1 , which is case 5,12 only
2- Leverage: average leverage in our model 1 = (K+1)/n, (3+1)/68 = 0.0588cases with 3
44
Multiple Liner Regression
Frist one with 14 case and the second one with 54 case then compare Square
Sample 1
45
Multiple Liner Regression
Sample 2 :
46
Multiple Liner Regression
Modal Summary 2:
R Square for sample 1 with [14 ] cases : have R Square 0.757 , adjusted 0.548
R Square for sample 1 with [54] cases : have R Square 0.621 , adjusted 0.572
The R Square full sample in the model 1 can show in page 8 , with R square 0.624 and R Adjusted
0.587
As we note that Sample [2] is more near to the model 1 , because it has large number of
observation 54 while sample 1 have only 14 observation and this make it 0.757 a little between
different from 0.621 , Maybe we need larger sample to make a strong cross-validation check
I Just make this section as a fast application for this method and to show how it apply
47
Multiple Liner Regression
5 Appendices
48
Multiple Liner Regression
49
Multiple Liner Regression
39 .00414 .02359 1.58025 1.07319 -.06072 -.05061 .03406 -.00211 .04930 .06656 .04960
40 .00319 .07894 5.28887 1.20824 .04196 -.02132 -.00383 .01864 .09687 -.05831 -.07140
41 .00207 .05313 3.55987 1.17712 -.08978 .02981 .07691 -.02045 -.00791 .00099 .10410
42 .00166 .02062 1.38135 1.12205 -.06710 .01601 .04276 .01322 .00172 .01608 .07602
43 .00014 .07952 5.32780 1.23813 .00279 -.00829 -.00473 .01383 .01225 -.01839 -.00445
44 .00020 .03234 2.16653 1.17423 -.01487 -.02114 .01061 .00437 -.00838 .00149 .01787
45 .00200 .10300 6.90090 1.25720 -.09762 .02883 .09747 -.00760 .04573 -.01718 .08950
46 .00532 .03030 2.02985 1.07298 .06689 -.05129 -.08048 -.03225 .03165 .04200 -.05883
47 .00000 .05314 3.56066 1.20438 .00012 -.00102 -.00057 .00123 .00043 .00009 -.00028
48 .00004 .18885 12.65286 1.40943 .00350 -.00756 .00222 -.00320 .00588 .00081 -.00695
49 .00137 .01545 1.03485 1.11715 -.03236 -.00723 .00633 .03988 -.02893 .02894 .04040
50 .00290 .04527 3.03292 1.15138 -.07696 .03317 .08000 -.03390 -.04122 -.04238 .10330
51 .00116 .04045 2.71032 1.16945 -.04467 .03070 .03600 -.00685 -.03255 -.01513 .06095
52 .01504 .09674 6.48182 1.14669 -.21589 .16041 .22961 -.09883 .07216 -.11718 .23272
53 .07472 .17385 11.64785 1.06339 -.55395 -.23474 .50208 .01523 -.08383 .35079 .53766
54 .00042 .04811 3.22362 1.19192 -.00797 -.02747 .01091 -.00028 .02103 -.01498 .00652
55 .00036 .07910 5.29974 1.23542 -.00785 -.02024 -.00302 .03923 -.00583 .02200 .00074
56 .00687 .02462 1.64943 1.01987 .15619 -.06248 -.12029 .03157 -.06349 -.00625 -.14992
57 .00022 .16033 10.74208 1.35974 .00164 -.02893 -.00357 .01113 -.00725 -.01183 .00306
58 .00249 .05697 3.81698 1.17833 .01596 -.00919 -.04061 .05214 -.05457 .08834 -.02076
59 .01012 .04501 3.01560 1.04931 -.01418 .01867 .02344 .13157 .14975 -.08761 -.04382
60 .01275 .04268 2.85928 1.00447 .08575 -.11786 -.11374 .20262 -.07803 -.04460 -.08120
61 .04674 .06237 4.17861 .76431 .12311 -.31207 -.10821 .37256 .19029 -.06223 -.22489
62 .03023 .06335 4.24451 .90855 -.11401 -.27798 .21738 .01923 -.17083 .05091 .10262
63 .02461 .17143 11.48592 1.26450 .00541 -.15093 -.05739 .35594 .06446 -.07628 -.06697
64 .02681 .03342 2.23894 .76002 .09741 -.27857 -.02610 .16583 -.11004 -.00386 -.12350
65 .00884 .04455 2.98477 1.06523 .05372 .09861 -.00889 .02986 .01904 -.07915 -.08344
66 .01239 .02170 1.45421 .89085 .07098 .06867 -.03627 .05534 -.06443 -.11314 -.06680
67 .02092 .03876 2.59716 .87518 .18214 .10130 -.18048 .11286 -.10699 -.03605 -.18265
68 .01724 .03862 2.58728 .92325 .07954 .18303 -.04189 .00220 -.07044 -.05912 -.08553
Total N 68 68 68 68 68 68 68 68 68 68 68
50
Multiple Liner Regression
51
Multiple Liner Regression
Case Summariesa
Standardized
Cook's Centered Mahalanobis DFBETA Standardized Standardized Standardized Standardized Standardized Standardized
Distance Leverage Value Distance COVRATIO Intercept DFBETA X1 DFBETA X2 DFBETA X3 DFBETA X4 DFBETA X5 DFBETA X6
52
Multiple Liner Regression
53
Multiple Liner Regression
54
Multiple Liner Regression
Case Summariesa
Centered Standardize Standardize Standardize Standardize Standardize Standardize Standardize
Cook's Leverage Mahalanobis COVRATI d DFBETA d DFBETA d DFBETA d DFBETA d DFBETA d DFBETA d DFBETA
Distance Value Distance O Intercept X1 X2 X3 X4 X5 X6
1 .00000 .08348 5.67640 1.16201 . .00008 -.00257 .00144 .00374 . .
2 .00335 .07854 5.34050 1.14446 . -.00148 .00546 -.03678 .10714 . .
3 .00160 .10276 6.98801 1.18287 . -.00302 -.03646 .01689 .07173 . .
4 .02312 .09550 6.49369 1.11430 . .05903 .22213 -.21113 -.16806 . .
5 .03233 .03762 2.55844 .89500 . .22652 -.28102 .08832 .16517 . .
6 .00249 .10780 7.33050 1.18755 . .00173 -.03734 .01026 .09295 . .
7 .12481 .09139 6.21440 .84867 . .39054 .31251 -.49212 -.21289 . .
8 .02538 .06790 4.61734 1.04628 . .09937 -.17112 .15182 -.16840 . .
9 .00663 .07617 5.17924 1.12981 . .05598 -.06958 .06560 -.10004 . .
10 .00911 .02429 1.65205 .99502 . .12432 .04387 -.12977 .01560 . .
11 .00033 .05026 3.41744 1.11965 . -.00886 .01519 -.01673 .02412 . .
12 .03018 .06073 4.12971 1.00724 . .02166 -.04497 .07872 -.26826 . .
13 .00046 .04663 3.17091 1.11451 . -.00369 -.01221 .00525 .03675 . .
14 .00186 .04246 2.88701 1.10060 . -.00790 -.01770 .00315 .07556 . .
15 .01405 .07828 5.32279 1.10841 . -.00105 -.03192 -.02744 .22207 . .
16 .00346 .09264 6.29949 1.16384 . -.02943 .05123 -.05602 .07780 . .
17 .04171 .04452 3.02749 .88539 . -.21401 .27502 -.21975 -.08425 . .
18 .04397 .11603 7.88996 1.10704 . .19512 -.38997 .22969 .12483 . .
19 .00057 .07044 4.79019 1.14357 . .00597 -.04573 .03852 .00312 . .
20 .00108 .02094 1.42410 1.07418 . .04003 -.03577 .00581 .00585 . .
21 .01644 .04829 3.28381 1.03115 . .16312 .01361 -.16311 .13780 . .
22 .14680 .08826 6.00189 .78436 . .39524 .18013 -.59643 .41162 . .
23 .00275 .01666 1.13318 1.03976 . .03145 -.03532 -.01527 .04084 . .
24 .01143 .13654 9.28493 1.21131 . .09923 -.19827 .11461 .07574 . .
25 .01328 .05452 3.70709 1.06294 . .08454 -.19698 .10312 .07582 . .
26 .00673 .05002 3.40154 1.08572 . .08021 .05863 -.13437 .05749 . .
27 .00305 .02796 1.90129 1.06694 . .02021 .02021 -.06906 .06147 . .
28 .00439 .08156 5.54593 1.14532 . -.06049 .12228 -.07491 -.03154 . .
29 .01041 .05611 3.81575 1.07977 . -.11577 .07061 -.01328 -.12388 . .
30 .00008 .02540 1.72686 1.09196 . -.00280 .01415 -.00869 -.00370 . .
31 .06681 .10240 6.96294 1.02210 . -.39423 -.18387 .26072 .09534 . .
32 .00001 .03451 2.34641 1.10302 . -.00187 .00477 -.00241 -.00207 . .
33 .00063 .01524 1.03657 1.07047 . .00116 -.00921 .01375 .01891 . .
34 .00086 .04378 2.97733 1.10856 . -.02355 .03992 -.03878 .00106 . .
35 .00043 .05665 3.85217 1.12695 . .01494 .02220 -.02910 .01859 . .
36 .00248 .10704 7.27898 1.18653 . .08687 -.01023 -.03540 .05219 . .
37 .05202 .06316 4.29482 .93288 . -.21680 .19765 -.22496 .12320 . .
38 .00048 .15825 10.76082 1.26444 . -.03999 .00618 .01399 -.01985 . .
39 .00571 .02903 1.97390 1.04541 . -.05237 -.01297 -.01827 .06505 . .
55
Multiple Liner Regression
56
Multiple Liner Regression
57
Multiple Liner Regression
Case Summariesa
Analog of
Cook's
influence DFBETA for DFBETA for
statistics Leverage value constant X1(1) DFBETA for X2 DFBETA for X3 DFBETA for X4 DFBETA for X5 DFBETA for X6
58
Multiple Liner Regression
59
Multiple Liner Regression
60
Multiple Liner Regression
Case Summariesa
61
Multiple Liner Regression
62