Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
A cross-sectional data set consists of a sample of individuals, households, firms, cities, states,
countries, or a variety of other units, taken at a given point in time. A time series data set consists of
observations on a variable or several variables over time. Some data sets have both cross-sectional and
time series features. For example, suppose that two cross-sectional household surveys are taken in the
United States, one in 1985 and one in 1990. In 1985, a random sample of households is surveyed for
variables such as income, savings, family size, and so on. In 1990, a new random sample of households
is taken using the same survey questions. To increase our sample size, we can form a pooled cross
section by combining the two years. A panel data (or longitudinal data) set consists of a time series
for each cross-sectional member in the data set. The key feature of panel data that distinguishes them
from a pooled cross section is that the same cross-sectional units (individuals, firms, or counties in the
preceding examples) are followed over a given time period.
2. Always keep in mind that regression does not necessarily imply causation. Causality
must be justified, or inferred, from the theory that underlies the phenomenon that
is tested empirically.
3. Regression analysis may have one of the following objectives:
To estimate the mean, or average, value of the dependent variable, given the values of the
independent variables.
To test hypotheses about the nature of the dependence—hypotheses suggested by the underlying
economic theory.
To predict, or forecast, the mean value of the dependent variable, given the value(s) of the independent
variable(s) beyond the sample range.
One or more of the preceding objectives combined.
4. the population regression line (PRL) gives the average, or mean, value of the dependent
variable corresponding to each value of the independent variable in the population
as a whole
5.
12. ei = the estimator of ui. We call ei the residual term, or simply the residual. Conceptually, it is
analogous
13. to ui and can be regarded as the estimator of the latter. It is introduced in the SRF for the same reasons
as ui was introduced in the PRF. Simply stated, ei represents the difference between the
actual Y values and their estimated values from the sample regression
15.
16. the lower the p value, the greater the evidence against the null hypothesis.so
higher the p value lower the chance of rejecting null hypothesis
Rajesh garg
18.
19. Correlation analysis measure degree of linear association between two variables.
20. Stochastic variables have probability distributions.
21. Linearity means linear in parameter.
22. No multi collinearity
23. Random sampling
24. Fixed x values or x values independent of the error term
25. Zero mean of disturbance that means no specification bias
26. Homoscadecity
27. No serial or auto correlation
28. N> parameters
29. Var(x)> 0 no outliners
30. Standard error of regression (sigma cap) = sqrt( sigma ei / n-2)
31. Cov (b1,b2) = - x bar* variance of b2
32. For linearity we need non stochastic x, for unbiasedness we need error term is zero, random sampling,
no multinearity , linearity in parameters.
33. Var b2= sigma sq by sigma x^2
34. For min var we need no auto correlation and homoscadecity , linearity and unbiasedness
35. If n increases var b2 tends to zero that means consistency
36. If ui are not normal still estimators are BLUE
37. For joint test set up r sq = 0 follow f statistic.
38. bxy * byx = r2
39. AM of bxy * byx ≥ r2
40. Regression y on x and x on y intersect at mean of x and y
41. If r=0 that means y on x and x on y are perpendicular
42. If r =1/ -1 that means y on x and x on y coincide or parallel.
43. R 1.23 means regression coefficient between x and (combine y and z) . multiple correlation coefficient
is always greater or equal than any total correlation coefficient ( r12, r13, r23)
44.
Rajesh garg
45.
46.
47.
Rajesh garg
48.
49. An important property of R2 is that the larger the number of explanatory variables in a model, the
higher the R2 will be. It would then seem that if we want to explain a substantial amount of the variation
in a dependent variable, we merely have to go on adding more explanatory variables! comparing the
R2 values of two models with the same dependent variable but with differing numbers
of explanatory variables is essentially like comparing apples and oranges.
50.
51. The common practice is to add variables as long as the adjusted R2 increases (even though its numerical
value may be smaller than the unadjusted R2). But when does adjusted R2 increase? It can be shown
that will increase if the (absolute t) value of the coefficient of the added
variable is larger than 1, where the t value is computed under the null hypothesis
that the population value of the said coefficient
a. is zero.
52. The adjusted and unadjusted R2s are identical only when the unadjusted R2 is equal to 1.
53.
54. The d.f. of the total sum of squares (TSS) are always n-1 regardless of the number of explanatory variables
included in the model.
55. The assumptions made by the classical linear regression model (CLRM) are not necessary to compute OLS
estimators.
56. In the two-variable PRF, b2 is likely to be a more accurate estimate of B2 if the disturbances ui follow the
normal distribution. This is not true
57. The PRF gives the mean value of the dependent variable corresponding to each value of the independent
variable. A linear regression model means a model linear in the parameters.
58. Comparing two models on the basis of goodness of fit, dependent variable and n should be same for
comparison.
59. Standardised value regression will tell us effect of independent variable on dependent variable in exact
proportion of slope coefficient.
60. If correlation btw Y and X2 is zero . partial correlation btw Y and X2 holding X3 constant may not to zero.
61. Polynomial regression do not violate multicollinearity. Example total cost function
62. Precision of OLS estimators is measured by their standard errors and good of fit by r2
63. OLS estimators have minimum variance in the entire class of unbiased estimators whether linear or not. Bit
for this normality is necessary.
64. Under normality maximum likelihood and OLS estimators converge.
65. The interpretation of CI is that 95 out of 100 cases , this interval will have true B. we can not say that the
probability is 95% that the interval contains true B, because now interval is fixed and no longer random the
prob for B lies is only 0 or 1. This definition is valid for regression coefficient.
66. to compare the values of two models, the dependent variable must be in the same
form.
67. For the linear-in-variable (LIV) model, the slope coefficient is constant but the elasticity coefficient is
variable, whereas for the log-log model, the elasticity is constant but the slope coefficient is variable.
68. for the linear model the elasticity coefficient changes from point to point and that for the log-linear
model it remains the same at all points the elasticity coefficient for the linear model is often computed
at the sample mean values of X and Y to obtain a measure of average elasticity. If we add the
elasticity coefficients,we obtain an economically important parameter, called the returns to scale
Rajesh garg
parameter, which gives the response of output to a proportional change in inputs. If the sum of the two
elasticity coefficients is 1, we have constant returns to scale (i.e., doubling the inputs simultaneously
doubles the output); if it is greater than 1,we have increasing returns to scale (i.e., doubling the inputs
simultaneously more than doubles the output); if it is less than 1, we have decreasing returns to scale
(i.e., doubling the inputs less than doubles the output).
70.
a. B2 is instantaneous rate of growth
71.
72.
Rajesh garg
automobile). this threshold level of income is at the level −(B2/B1). (2) There is a satiety level of
consumption beyond which the consumer will not go no matter how high the income (even millionaires
do not generally own more than two or three cars at a time). This level is nothing but the asymptote B1
For such commodities, the reciprocal model of this figure is the most appropriate.
77. Figure (c) is the celebrated Phillips curve of macroeconomics. Based on the British data on the percent
rate of change of money wages (Y) and the unemployment rate (X) in percent, Phillips obtained a curve
similar to Figure (c).20 As this figure shows, there is asymmetry in the response of wage changes to the
level of unemployment. Wages rise faster for a unit change in unemployment if the unemployment rate
is below UN, which is called the natural rate of unemployment by economists, than they fall for an
equivalent change when the unemployment rate is above the natural level, B1 indicating the asymptotic
floor for wage change. This particular feature of the Phillips curve may be due to institutional factors,
such as union bargaining power, minimum wages, or unemployment insurance.
80.
81. First, in the model without the intercept, we use raw sums of squares and cross products, whereas in
the intercept-present model, we use mean-adjusted sums of squares and cross products.
82. Second, the d.f. in computing sigma cap sq is n-1 now rather than n-2 , since we have only one
unknown.
83. Third, the conventionally computed r2 formula we have used thus far explicitly assumes that the
model has an intercept term. Therefore, you should not use that formula. If you use it, sometimes you
will get nonsensical results because the computed r2 may turn out to be negative.
84. Finally, for the model that includes the intercept, the sum of the estimated residuals, is always zero,
but this need not be the case for a model without the intercept term. For all these reasons, one may use
the zero-intercept model only if there is strong theoretical reason for it, as in Okun’s law or some areas
of economics and finance.
85. If X are multiplied by w and Y are multiplied by v, b2 factor by v/w and b1 factor by v only.sigma cap sq by
v2 Keep following things in mind while scaling, First, the r2 and t value are same, Second , when Y
and X are measured in the same units of measurement the slope coefficients as well as their standard
errors remain the same although the intercept values and their standard errors are different Third,
when the Y and X variables are measured in different units of measurement, the slope coefficients are
different, but the interpretation does not change.
86. Standardized regression is regression from origin.
87.
88. The interpretation is that if the (standardized) regressor increases by one standard deviation, the
average value B2* of the (standardized) regressand increases by standard deviation units. you will see
the advantage of using standardized variables, for standardization puts all variables on equal footing
because all standardized variables have zero means and unit variances. This will help us to find which
Rajesh garg
regressor has more effect on regrassand by looking at their coefficient. Higher the coefficient higher
the effect.
89.
90. Log inverse model is basically based on microeconomics return to factor diagram i.e. increasing at
increasing rate then increasing at decreasing rate.
91. In log log model ui follow log normal distribution with mean e^(siqma2 /2) and variance {e^(sigma sq.) *
[e^(sigma sq.) - 1]
93. Interpretation of R2 or ESS, how much variation in dependent variable is explained by variation in
independent variable or variation percentage explained by regression.
94. Regression models that contain only dummy explanatory variables are called analysis-of-variance
(ANOVA) models. 1Since dummy variables generally take on values of 1 or 0, they are nonstochastic;
that is, their values are fixed. And since we have assumed all along that our X variables are fixed in
repeated sampling, the fact that one or more of these X variables are dummies does not create any
special problems insofar as estimation of model is concerned. In short, dummy explanatory variables
do not pose any new estimation problems and we can use the customary OLS method to estimate the
parameters of models that contain dummy explanatory variables. a regression on an intercept and a
dummy variable is a simple way of finding out if the mean values of two groups differ. If the dummy
coefficient B2 is statistically significant (at the chosen level of significance level), we say that the two
means are statistically different. If it is not statistically significant, we say that the two means are not
statistically significant.
95. The general rule is: If a model has the common intercept, B1, and if a qualitative
variable has m categories, introduce only (m - 1) dummy variables. If this rule is not
followed, we will fall into what is known as the dummy variable trap, that is, the situation of perfect
collinearity or multicollinearity, if there is more than one perfect relationship among the variables.
Another way to resolve the perfect collinearity problem is to keep as many dummies as the number of
categories but to drop the common intercept term, B1, from the model; that is, run the regression
through the origin.
96. Regression models containing a combination of quantitative and qualitative variables are called
analysis-of-covariance (ANCOVA) models.
97. the two regression lines differ in their intercepts but their slopes are
the same. In other words, these two regression lines are parallel.
98.
Rajesh garg
100. number of dummies for each qualitative variable is one less than the number
of categories of that variable.
101. To comparing two or more regressions is popularly known as the Chow test, this is
Rajesh garg
118. Difficulty in assessing the individual contributions of explanatory
variables to the explained sum of squares (ESS) or R2.
119. If the goal of the study is to use the model to predict or forecast the future mean value of the dependent
variable, collinearity per se may not be bad. On the other hand, if the objective of the study is not only
prediction but also reliable estimation of the individual parameters of the chosen model, then serious
collinearity may be bad, because we have seen that it leads to large standard errors of the estimators.
Rajesh garg
136. heteroscadicity may resukt because of outliners in the sample, omission of variables and
skewness of distribution of regressors in the model. Wrong functional form, incorrect transformation of
variables
137. CONSEQUENCES OF HETEROSCEDASTICITY
138. OLS estimators are still linear. They are still unbiased. But they no longer have minimum variance; that
is, they are no longer Efficient. In short, OLS estimators are no longer BLUE in small as well as in
large samples (i.e., asymptotically). It might give consistent estimators
139. The usual formulas to estimate the variances of OLS estimators are generally biased. A priori we
cannot tell whether the bias will be positive (upward bias) or negative (downward bias). A positive bias
occurs if OLS overestimates the true variances of estimators, and a negative bias occurs if OLS
underestimates the true variances of estimators. The bias arises from the fact that , the conventional
estimator of true standard error is no longer an unbiased .
140. in the presence of heteroscedasticity, the usual hypothesis-testing routine is not
reliable, raising the possibility of drawing misleading conclusions.
141. DETECTION OF HETEROSCEDASTICITY
Rajesh garg
b. We first estimate regression by OLS, obtaining the residuals, ei.
c. We then run the following auxiliary regression: The term vi is the residual term in the
auxiliary regression.
d. Obtain the R2 value from the auxiliary regression. Under the null hypothesis that there is no
heteroscedasticity. .
148. This test is also used for specification bias
149. Glejser Test
a. Glejser, has maintained that in large samples the preceding models are fairly good in
detecting heteroscedasticity. Therefore, Glejser’s test can be used as a diagnostic tool in large
samples. The null hypothesis in each case is that there is no heteroscedasticity; that is, B2 = 0.
If this hypothesis is rejected, there is probably evidence of heteroscedasticity
b.
150. Goldfeld-Quandt test
a. Order the Xi, omit c central observation and split remaining n-c observation into two part. Fit
regression, find RSS of both, obtain $= RSS2 / RSS1. F statistic with (n-c-2k)/2 df.
151. Note: There is no general test of heteroscedasticity that is free of any assumption about which variable
the error term is correlated with.
a.
154. When True sigma2i Is Unknown
a. Case 1: The Error Variance Is Proportional to Xi: The Square Root Transformation
b.
i. It is important to note that to estimate we must use the regression-
through-the-origin estimating procedure
c. Case 2: Error Variance Proportional to Xi^2
d.
155. Respecification of the Model
Rajesh garg
a. Instead of speculating about , sometimes a respecification of the PRF—choosing a different
functional form can reduce heteroscedasticity.
156. However, if the sample size is reasonably large, we can use White’s procedure to obtain
heteroscedasticity corrected standard errors. “Robust standard error”
157. The term autocorrelation can be defined as “correlation between members of observations ordered
in time (as in time series data) or space (as in crosssectional data). Just as heteroscedasticity is
generally associated with cross-sectional data autocorrelation is usually associated with time series data
(i.e., data ordered in temporal sequence), although, as the preceding definition suggests, autocorrelation
can occur in cross-sectional data also, in which case it is called spatial correlation (i.e., correlation in
space rather than in time).
Rajesh garg
i. it is very important to note the assumptions underlying the d
statistic:
c. The regression model includes an intercept term. Therefore, it cannot be used to determine
autocorrelation in models of regression through the origin. Ui is normally distributed .
d. The X variables are nonstochastic; that is, their values are fixed in repeated sampling. No
missing term in time series data.
e. The disturbances ut are generated by the following mechanism
163.
i.
164. The Breusch-Godfrey (BG) Test:
Rajesh garg
a. This test is general in that it allows for (1) stochastic regressors, such as the lagged values of
the dependent variables, (2) higher-order autoregressive schemes, such as AR(1), AR(2), etc.,
and (3) simple or higher-order moving averages of the purely random error terms.
b. Run the regression ,obtain residual from this regression, . Now run the following regression:
i. That is, regress the residual at time t on the original regressors, including the
intercept and the lagged values of the residuals up to time .Obtain the value of R2 of
this regression. This is called the auxiliary regression. Under the null hypothesis
that all the coefficients of the lagged residual terms are simultaneously equal to zero,
it can be shown that in large samples That is, in large samples, the product of the
sample size and follows the chi-square distribution with k degrees of freedom (i.e.,
the number of lagged residual terms). In econometrics literature, the BG test is
known as the Lagrange multiplier test. But keep in mind that the
BG test is of general applicability, whereas the Durbin-Watson test assumes only
first-order serial correlation.
c.
a.
Rajesh garg
b. apply OLS to the transformed
variables Y* and X*, the estimators thus obtained will have the desirable BLUE property
generalized difference equations In this differencing procedure we lose one observation
because the first sample observation has no antecedent. To avoid this loss of one observation,
a. TRUE MODEL
b. ESTIMATED MODEL
c. CONSEQUENCES OF OMISSION:
180. If X3 is correlated to X2, A1 & A2 are biased and inconsistent
181. If X2 and X3 are uncorrelated,a2 is unbiased. It is consistent as well. But a1 still remains biased, unless
mean of X3 is zero.
182. Error variance is incorrectly estimated and biased.
183. estimated variance of b2 is a biased estimator of the variance of the true estimator b2. Even in the case
X2 and X3 are uncorrelated, this variance remains biased. As a result, the usual confidence interval and
hypothesis-testing procedures are unreliable.
a. ESTIMATED MODEL
b. TRUE MODEL
c. CONSEQUENCES OF INCLUSION:
Rajesh garg
186. The OLS estimators of the “incorrect” model are unbiased (as well as consistent).
187. Error variance is correctly estimated. As a result, the usual confidence interval and hypothesis-testing
procedures are reliable and valid.
188. However, the a’s estimated from the regression are inefficient— their variances will be generally
larger than those of the b’s estimated from the true model In short, the OLS estimators are LUE
(linear unbiased estimators) but not BLUE.
Rajesh garg