Lessons Plans Flashcards

1
Q

Building and Estimating a Model (5 step test)

A
  1. ) Hypothesize a relation between X variables and Y
  2. ) Use sample data to estimate the unknown parameters in the model
  3. ) Compute the standard errors of the parameters
  4. ) Statistically evaluate the usefulness of the model
  5. ) Use the model for prediction
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

When do we say there is heteroskedasticity in the data?

A

The variance of e changes when X changes (the variance of e is not constant and does not equal variance^2)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the assumptions about the error term?

A
  1. ) E[e] = 0
  2. ) the variance of e does not change when X changes
  3. ) the probability of e is normal. (Generally, this is difficult to know whether it’s not true)
  4. ) The value of ei is independent of ei for any two observations I and j. Example, if any two of you of you in the room work for the same company, your errors in the regression of your income on your parent’s income would likely be correlated. `
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Test whether B1 is “significantly” different from zero

A

1.) Estimate variance^2
2.) a = compute a confidence interval around Bhat1
b = compute the T-statistic for the difference between Bhat1 and 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Total Sum of Squares (TSS)

A

the sum of squares of the deviations from the mean of the dependent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Explained Sum of Squares

A

the sum of squares of the deviations from the mean of the predictions (i.e. predictions of the dependent variable Y in this case)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Residual Sum of Squares (RSS)

A

the sum of squares of the residuals (i.e. the errors in the predictions)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How is the Total Sum of Squares (TSS) partitioned into two parts?

A

This partitioning is called variance decomposition. We fit a regression line and see how much of the variation can be explained by the fitted relationship (the explained sum of squares, ESS) and how much is left over (the residual sum of squares, RSS).

the smaller the RSS, relative to TSS, the better the regression line appears to fit.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

R^2is defined as the …

A

fraction of the total variability in Y that is explained by the regression model. R^2 is measure of “goodness of fit” and (rarely) called the “coefficient of determination.”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

R^2 =

A

Variation in Y that is explained by X/Total variation in y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

OLS (Ordinary Least Squares, also known as minimizing squared errors, which is what we’ve been doing) gives us

A

Betas that minimize the RSS (keeps squared residuals as small as possible), thus gives largest possible R^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Adjusted R^2

A

adjusts for the number of independent variables (i.e. predictors) in the regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

When moving to multivariate regressions the interpretation of R^2 will be…

A

“what fraction of the variation in Y can be explained by the X variables.”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are the two ways to use a bivariate prediction equation?

A
  1. ) To estimate the mean value of Y for a population with a given X.
  2. ) To estimate the value of Y for an individual with a given X.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

The smaller the RSS is, relative to TSS…

A

the better the regression line appears to fit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How does the total sum of squares partition into two parts? (Variance Decomposition)

A
  1. ) We fit a regression line and see how much of the variation can be explained by the fitted relationship (the explained sum of squares, ESS)
  2. ) How much is left over (the residual sum of squares, RSS)
17
Q

OLS gives us…

A

Betas that minimize the RSS, thus giving the largest possible R^2

18
Q

“Adjusted R squared”

A

adjusts for the number of independent variables in the regression

19
Q

when there is only one X variable explain Y then r =

A

square root of r squared

20
Q

What are the two ways we would use a prediction equation?

A
  1. ) To estimate the mean value of Y for a population with a given X
  2. ) To estimate the value of Y for an individual with a given X
21
Q

We should be less confident with our answer for an individual than for a …

A

group average

22
Q

Why should we include more than one variable in our regressions?

A
  1. ) To improve the precision of our predictions of Y. By lowering the standard error of the regression model, we also lower the standard errors of our coefficients and thus have more precisely estimated coefficients.
  2. ) To avoid biased estimates of the coefficients (“omitted variable bias”)
23
Q

If N is large and k is small, r squared and adj r squared will be…

A

very similar

24
Q

If N is small or k is large, r squared and adj r squared will be …

A

very different. In these cases you should use Adjusted R Squared

25
Q

It is impossible to estimate a model where…

A

k > N. This is because there are not enough degrees of freedom in the dat to estimate all of the k parameters.

26
Q

If we don’t pass the F test…

A

we should be very concerned with the validity of our model or its predictions