代写*OverviewRegressionAnalysiswithCrossSectionalData
100%原创包过,高质代写&免费提供Turnitin报告--24小时客服QQ&微信:120591129
代写

OverviewRegressionAnalysiswithCrossSectionalData
Definition of the multiple linear regression model
Motivation for multiple regression

Incorporate more explanatory factors into the model

Explicitly hold fixed other factors that otherwise would be in

Allow for more flexible functional forms

Example: Wage equation
Interpretation of the multiple regression model

The multiple linear regression model manages to hold the values of other explanatory variables fixed even if, in reality, they are correlated with the explanatory variable under consideration

„Ceteris paribus“-interpretation

It has still to be assumed that unobserved factors do not change if the explanatory variables are changed

Example: Determinants of college GPA

Interpretation
Holding ACT fixed, another point on high school grade point average is associated with another .453 points college grade point average

Or: If we compare two students
with the same ACT, but the hsGPA of student A is one point higher, we predict student A to have a colGPA that is .453 higher than that of student B
Holding high school grade point average fixed, another 10 points on ACT are associated with less than one point on college GPA
Standard assumptions for the multiple regression model

Assumption MLR.1 (Linear in parameters)

Assumption MLR.2 (Random sampling)
Standard assumptions for the multiple regression model (cont.)

Assumption MLR.3 (No perfect collinearity)

Remarks on MLR.3

The assumption only rules out
perfect collinearity/correlation bet-ween explanatory variables; imperfect correlation is allowed

If an explanatory variable is a perfect linear combination of other explanatory variables it is superfluous and may be eliminated

Constant variables are also ruled out (collinear with intercept)

Example for perfect collinearity: small sample

Example for perfect collinearity: relationships between regressors
Standard assumptions for the multiple regression model (cont.)

Assumption MLR.4 (Zero conditional mean)

In a multiple regression model, the zero conditional mean assumption is much more likely to hold because fewer things end up in the error

Example: Average test scores

Discussion of the zero mean conditional assumption

Explanatory variables that are correlated with the error term are called
endogenous; endogeneity is a violation of assumption MLR.4

Explanatory variables that are uncorrelated with the error term are called
exogenous; MLR.4 holds if all explanat. var. are exogenous

Exogeneity is the key assumption for a causal interpretation of the regression, and for unbiasedness of the OLS estimators
Theorem 3.1 (Unbiasedness of OLS)

Unbiasedness is an average property in repeated samples; in a given sample, the estimates may still be far away from the true values
Including irrelevant variables in a regression model
Omitting relevant variables: the simple case;
Conclusion: All estimated coefficients will be biased
Standard assumptions for the multiple regression model (cont.)

Assumption MLR.5 (Homoscedasticity)

Example: Wage equation

Short hand notation
Assumption MLR.6 (Normality of error terms)
代写

OverviewRegressionAnalysiswithCrossSectionalData
Theorem 3.2 (Sampling variances of OLS slope estimators)

An example for multicollinearity

Discussion of the multicollinearity problem

In the above example, it would probably be better to lump all expen-diture categories together because effects cannot be disentangled

In other cases, dropping some independent variables may reduce multicollinearity (but this may lead to omitted variable bias)

Only the sampling variance of the variables involved in multicollinearity will be inflated; the estimates of other effects may be very precise

Note that multicollinearity is not a violation of MLR.3 in the strict sense

Multicollinearity may be detected through „variance inflation factors“
Estimating the error variance
Theorem 3.3 (Unbiased estimator of the error variance)
Efficiency of OLS: The Gauss-Markov Theorem

Under assumptions MLR.1 - MLR.5, OLS is unbiased

However, under these assumptions there may be many other estimators that are unbiased

Which one is the unbiased estimator with the
smallest variance?

In order to answer this question one usually limits oneself to linear estimators, i.e. estimators linear in the dependent variable
Theorem 3.4 (Gauss-Markov Theorem)

Under assumptions MLR.1 - MLR.5, the OLS estimators are the best linear unbiased estimators (BLUEs) of the regression coefficients, i.e.

OLS is only the best estimator if MLR.1 – MLR.5 hold; if there is heteroscedasticity for example, there are better estimators.
Estimation of the sampling variances of the OLS estimators

Note that these formulas are only valid under assumptions MLR.1-MLR.5 (in particular, there has to be homoscedasticity)

Terminology

Theorem 4.1 (Normal sampling distributions)
Testing hypotheses about a single population parameter

Theorem 4.1 (t-distribution for standardized estimators)

Null hypothesis (for more general hypotheses, see below)

t-statistic (or t-ratio)

Distribution of the t-statistic
if the null hypothesis is true
Goal: Define a rejection rule so that, if it is true, H0 is rejected only with a small probability (= significance level, e.g. 5%)

Testing against one-sided alternatives (greater than zero)

Example: Wage equation

Test whether, after controlling for education and tenure, higher work experience leads to higher hourly wages

Example: Wage equation (cont.)

Testing against one-sided alternatives (less than zero)

Example: Student performance and school size

Test whether smaller school size leads to better student performance

Example: Student performance and school size (cont.)

Example: Student performance and school size (cont.)

Alternative specification of functional form:

Example: Student performance and school size (cont.)

Testing against two-sided alternatives

Example: Determinants of college GPA

„Statistically significant“ variables in a regression

If a regression coefficient is different from zero in a two-sided test, the corresponding variable is said to be
„statistically significant“

If the number of degrees of freedom is large enough so that the nor-mal approximation applies, the following rules of thumb apply:

Guidelines for discussing economic and statistical significance

If a variable is statistically significant, discuss the magnitude of the coefficient to get an idea of its economic or practical importance
The fact that a coefficient is statistically significant does not necessa-rily mean it is economically or practically significant!

If a variable is statistically and economically important but has the „wrong“ sign, the regression model might be misspecified

If a variable is statistically insignificant at the usual levels (10%, 5%, 1%), one may think of dropping it from the regression

If the sample size is small, effects might be imprecisely estimated so that the case for dropping insignificant variables is less strong
Testing more general hypotheses about a regression coefficient

Null hypothesis

t-statistic
The test works exactly as before, except that the hypothesized value is substracted from the estimate when forming the statistic

Example: Campus crime and enrollment

An interesting hypothesis is whether crime increases by one percent if enrollment is increased by one percent
Computing p-values for t-tests

If the significance level is made smaller and smaller, there will be a point where the null hypothesis cannot be rejected anymore

The reason is that, by lowering the significance level, one wants to avoid more and more to make the error of rejecting a correct H0

The smallest significance level at which the null hypothesis is still rejected, is called the
p-value of the hypothesis test

A small p-value is evidence against the null hypothesis because one would reject the null hypothesis even at small significance levels

A large p-value is evidence in favor of the null hypothesis

P-values are more informative than tests at fixed significance levels

How the p-value is computed (here: two-sided test)
Confidence intervals

Simple manipulation of the result in Theorem 4.2 implies that

Interpretation of the confidence interval

The bounds of the interval are random

In repeated samples, the interval that is constructed in the above way will cover the population regression coefficient in 95% of the cases

Confidence intervals for typical confidence levels

Relationship between confidence intervals and hypotheses tests

Example: Model of firms‘ R&D expenditures
Testing hypotheses about a linear combination of parameters

Example: Return to education at 2 year vs. at 4 year colleges

Impossible to compute with standard regression output because

Alternative method

Estimation results

This method works
always for single linear hypotheses
Testing multiple linear restrictions: The F-test

Testing exclusion restrictions

Estimation of the unrestricted model

Estimation of the restricted model

Test statistic

Rejection rule (Figure 4.7)

Test decision in example

Discussion

The three variables are „jointly significant“

They were not significant when tested individually

The likely reason is multicollinearity between them
Test of overall significance of a regression

The test of overall significance is reported in most regression packages; the null hypothesis is usually overwhelmingly rejected
Testing general linear restrictions with the F-test

Example: Test whether house price assessments are rational

Unrestricted regression

Restricted regression

Test statistic

Regression output for the unrestricted regression

The F-test works for general multiple linear hypotheses
For all tests and confidence intervals, validity of assumptions MLR.1 – MLR.6 has been assumed. Tests may be invalid otherwise.
Models with interaction terms

Interaction effects complicate interpretation of parameters

Reparametrization of interaction effects

Advantages of reparametrization

Easy interpretation of all parameters

Standard errors for partial effects at the mean values available

If necessary, interaction may be centered at other interesting values
Qualitative Information

Examples: gender, race, industry, region, rating grade, …

A way to incorporate qualitative information is to use dummy variables

They may appear as the dependent or as independent variables
A single dummy independent variable

Dummy variable trap

Estimated wage equation with intercept shift

Does that mean that women are discriminated against?

Not necessarily. Being female may be correlated with other produc-tivity characteristics that have not been controlled for.

Using dummy explanatory variables in equations for log(y)
Using dummy variables for multiple categories

1) Define membership in each category by a dummy variable

2) Leave out one category (which becomes the base category)
代写

OverviewRegressionAnalysiswithCrossSectionalData