Statistics "Fill in the blank spaces" quiz

ClearerKoala avatar
ClearerKoala
·
·
Download

Start Quiz

Study Flashcards

60 Questions

  • Linear regression is used to explain the relationship between a dependent variable and one or more ______ variables.

independent

  • The dependent variable is a ______ variable.

continuous

  • The simple linear regression model has ______ independent variable(s).

one

  • The multiple linear regression model has ______ or more independent variables.

two

  • The linear regression model describes how the dependent variable is related to the independent variables and the ______ term.

error

  • The estimated regression equation shows how to calculate predicted values of the dependent variable using the values of the ______ variables.

independent

  • The least squares method is used to calculate the coefficients so that the errors are as ______ as possible.

small

  • The coefficient of determination (R-squared) provides a measure of the ______ of fit for the estimated regression equation.

goodness

  • Adjusted R-squared corrects for the number of independent variables and is ______ to R-squared.

preferred

  • The t-test is used to determine whether the relationship between the dependent variable and ______ independent variable is significant.

one

  • The ANOVA table shows the total variation, explained variation due to regression, and unexplained variation due to ______.

error

  • The goal is to find a regression model with coefficients that are ______ significant.

jointly

  • An estimator is consistent if it converges in probability to the population parameter as the sample size increases. The probability that the estimator obtained from a sample size will be arbitrarily close to the population parameter goes to 1 as the sample size increases.

true

  • The OLS estimator is unbiased under assumptions 1-4 (with the zero conditional mean assumption).

true

  • Under assumptions 1-4’ (with the assumption that the regressors are uncorrelated with the error term), the OLS estimator is ______.

consistent

  • Unbiasedness is ideal but if it cannot be achieved in a small sample, then ______ can be achieved with a large sample.

consistency

  • Omitted variable bias occurs when a relevant variable is omitted from the model and the coefficient will be biased if the omitted variable and the included variable are ______.

correlated

  • Under assumptions 1-5 (Gauss Markov assumptions), the coefficients have asymptotically ______ sampling distribution.

normal

  • In large samples, the normality assumption is not always needed for the OLS estimators to be normal and the t-tests and F-tests to be ______.

valid

  • OLS properties hold for any sample, including expected values and unbiasedness under assumptions 1-4 and variance formulas under ______.

assumptions 1-5

  • Gauss-Markov theorem (BLUE) holds under ______ 1-5.

assumptions

  • As the sample size increases, standard errors change at a rate of 1/sample size and with larger sample size, standard errors are ______, leading to more significance of the coefficients.

lower

  • The OLS estimator is consistent if the omitted variable is ______ or uncorrelated.

irrelevant

  • The OLS estimator is ______ if assumptions 1-4’ hold.

consistent

  • Linear regression models the relationship between a dependent variable and one or more ______ variables.

independent

  • The regression model can be in linear or ______-linear form, and taking logs of variables changes the interpretation of coefficients.

log

  • Gauss Markov assumptions are standard assumptions for the linear regression model, including linearity in parameters, random sampling, no perfect collinearity, ______, and homoscedasticity.

exogeneity

  • [Blank] means the variance of the error term is constant for each independent variable while heteroscedasticity means the variance differs.

homoscedasticity

  • The unbiasedness of the OLS estimators is derived from ______ Markov assumptions.

Gauss

  • The standard errors measure how precisely the regression coefficients are calculated, and lower variance in error term and higher variance in ______ variable is desirable.

independent

  • The variance of the error term can be ______, and the variances of the OLS estimators depend on it.

estimated

  • The sample variability in OLS coefficients depends on the variances of the error term and ______ variable.

independent

  • The coefficients are random as the sample is random, and the expected values of the sample coefficients are the ______ parameters.

population

  • The relationship between y and x is ______ in the population, but the regression model can have logged, squared, or interaction variables.

linear

  • [Blank] or zero conditional mean implies the expected value of the error term given independent variable x is zero.

Exogeneity

  • [Blank] is when the variance of the error term is constant for each independent variable.

Homoscedasticity

  • Linear regression models the relationship between a dependent variable and one or more ______ variables.

independent

  • The ______ model can be in linear or log-linear form, and taking logs of variables changes the interpretation of coefficients.

regression

  • Gauss Markov assumptions are standard assumptions for the linear regression model, including linearity in parameters, ______ sampling, no perfect collinearity, exogeneity, and homoscedasticity.

random

  • ______ means the variance of the error term is constant for each independent variable while heteroscedasticity means the variance differs.

Homoscedasticity

  • The unbiasedness of the OLS estimators is derived from ______ assumptions.

Gauss Markov

  • The standard errors measure how precisely the regression coefficients are calculated, and lower variance in error term and higher variance in independent variable is ______.

desirable

  • The variance of the error term can be estimated, and the variances of the OLS estimators ______ on it.

depend

  • The sample variability in OLS coefficients depends on the variances of the error term and ______ variable.

independent

  • The coefficients are random as the sample is random, and the expected values of the sample coefficients are the ______ parameters.

population

  • The relationship between y and x is ______ in the population, but the regression model can have logged, squared, or interaction variables.

linear

  • Exogeneity or zero conditional mean implies the expected value of the error term given independent variable x is ______.

zero

  • ______ is when the variance of the error term is constant for each independent variable.

Homoscedasticity

  • Heteroscedasticity refers to a scenario where the variance of the error term differs with the ______ variables.

independent

  • Under heteroscedasticity, OLS estimators are still unbiased and consistent, but the variance formulas for the OLS estimators are not ______.

valid

  • The t-tests and F-tests are not valid under heteroscedasticity, and the OLS estimator is not the best linear unbiased estimator (BLUE).

  • Hypothesis testing for heteroscedasticity involves testing whether the expected value of the error term varies with the ______ variables.

independent

  • The Breusch-Pagan test, White test, and Alternative White test are commonly used tests for ______.

heteroscedasticity

  • Robust standard errors should be used when ______ is found.

heteroscedasticity

  • Weighted Least Squares (WLS) can be used to estimate the model if the heteroskedasticity form is ______.

known

  • Feasible Generalized Least Squares (FGLS) transforms the variables to get homoscedasticity if the heteroscedasticity form is ______ known.

not

  • The R-squared for the regressions of squared residuals on independent variables is used to calculate the test statistics for heteroscedasticity ______.

tests

  • The F-test and LM-test are commonly used tests for overall significance of ______.

heteroscedasticity

  • The regression model for price needs correction for ______.

heteroscedasticity

  • The R-squared for the regressions of squared residuals on independent variables is used to calculate the test statistics for heteroscedasticity tests for ______ price.

log

Study Notes

Linear Regression Overview

  • Linear regression is used to explain the relationship between a dependent variable and one or more independent variables.
  • The dependent variable is a continuous variable, while the independent variables can be continuous, discrete, or indicator variables.
  • The simple linear regression model has one independent variable, while the multiple linear regression model has two or more independent variables.
  • The linear regression model describes how the dependent variable is related to the independent variables and the error term.
  • The estimated regression equation shows how to calculate predicted values of the dependent variable using the values of the independent variables.
  • The least squares method is used to calculate the coefficients so that the errors are as small as possible.
  • The coefficient of determination (R-squared) provides a measure of the goodness of fit for the estimated regression equation.
  • Adjusted R-squared corrects for the number of independent variables and is preferred to R-squared.
  • The t-test is used to determine whether the relationship between the dependent variable and one independent variable is significant.
  • The F-test is used to test whether the relationship between the dependent variable and all independent variables is significant.
  • The ANOVA table shows the total variation, explained variation due to regression, and unexplained variation due to error.
  • The goal is to find a regression model with coefficients that are jointly significant.

Linear Regression Overview

  • Linear regression is used to explain the relationship between a dependent variable and one or more independent variables.
  • The dependent variable is a continuous variable, while the independent variables can be continuous, discrete, or indicator variables.
  • The simple linear regression model has one independent variable, while the multiple linear regression model has two or more independent variables.
  • The linear regression model describes how the dependent variable is related to the independent variables and the error term.
  • The estimated regression equation shows how to calculate predicted values of the dependent variable using the values of the independent variables.
  • The least squares method is used to calculate the coefficients so that the errors are as small as possible.
  • The coefficient of determination (R-squared) provides a measure of the goodness of fit for the estimated regression equation.
  • Adjusted R-squared corrects for the number of independent variables and is preferred to R-squared.
  • The t-test is used to determine whether the relationship between the dependent variable and one independent variable is significant.
  • The F-test is used to test whether the relationship between the dependent variable and all independent variables is significant.
  • The ANOVA table shows the total variation, explained variation due to regression, and unexplained variation due to error.
  • The goal is to find a regression model with coefficients that are jointly significant.

OLS Asymptotics

  • An estimator is consistent if it converges in probability to the population parameter as the sample size increases.
  • The probability that the estimator obtained from a sample size will be arbitrarily close to the population parameter goes to 1 as the sample size increases.
  • The OLS estimator is unbiased under assumptions 1-4 (with the zero conditional mean assumption).
  • Under assumptions 1-4’ (with the assumption that the regressors are uncorrelated with the error term), the OLS estimator is consistent.
  • Unbiasedness is ideal but if it cannot be achieved in a small sample, then consistency can be achieved with a large sample.
  • Omitted variable bias occurs when a relevant variable is omitted from the model and the coefficient will be biased if the omitted variable and the included variable are correlated.
  • The OLS estimator is consistent if the omitted variable is irrelevant or uncorrelated.
  • Under assumptions 1-5 (Gauss Markov assumptions), the coefficients have asymptotically normal sampling distribution.
  • In large samples, the normality assumption is not always needed for the OLS estimators to be normal and the t-tests and F-tests to be valid.
  • OLS properties hold for any sample, including expected values and unbiasedness under assumptions 1-4 and variance formulas under assumptions 1-5.
  • Gauss-Markov theorem (BLUE) holds under assumptions 1-5.
  • As the sample size increases, standard errors change at a rate of 1/sample size and with larger sample size, standard errors are lower, leading to more significance of the coefficients.

Introduction to Linear Regression

  • Linear regression models the relationship between a dependent variable and one or more independent variables.
  • The regression model can be in linear or log-linear form, and taking logs of variables changes the interpretation of coefficients.
  • Gauss Markov assumptions are standard assumptions for the linear regression model, including linearity in parameters, random sampling, no perfect collinearity, exogeneity, and homoscedasticity.
  • Homoscedasticity means the variance of the error term is constant for each independent variable while heteroscedasticity means the variance differs.
  • The unbiasedness of the OLS estimators is derived from Gauss Markov assumptions.
  • The standard errors measure how precisely the regression coefficients are calculated, and lower variance in error term and higher variance in independent variable is desirable.
  • The variance of the error term can be estimated, and the variances of the OLS estimators depend on it.
  • The sample variability in OLS coefficients depends on the variances of the error term and independent variable.
  • The coefficients are random as the sample is random, and the expected values of the sample coefficients are the population parameters.
  • The relationship between y and x is linear in the population, but the regression model can have logged, squared, or interaction variables.
  • Exogeneity or zero conditional mean implies the expected value of the error term given independent variable x is zero.
  • Homoscedasticity is when the variance of the error term is constant for each independent variable.

Introduction to Linear Regression

  • Linear regression models the relationship between a dependent variable and one or more independent variables.
  • The regression model can be in linear or log-linear form, and taking logs of variables changes the interpretation of coefficients.
  • Gauss Markov assumptions are standard assumptions for the linear regression model, including linearity in parameters, random sampling, no perfect collinearity, exogeneity, and homoscedasticity.
  • Homoscedasticity means the variance of the error term is constant for each independent variable while heteroscedasticity means the variance differs.
  • The unbiasedness of the OLS estimators is derived from Gauss Markov assumptions.
  • The standard errors measure how precisely the regression coefficients are calculated, and lower variance in error term and higher variance in independent variable is desirable.
  • The variance of the error term can be estimated, and the variances of the OLS estimators depend on it.
  • The sample variability in OLS coefficients depends on the variances of the error term and independent variable.
  • The coefficients are random as the sample is random, and the expected values of the sample coefficients are the population parameters.
  • The relationship between y and x is linear in the population, but the regression model can have logged, squared, or interaction variables.
  • Exogeneity or zero conditional mean implies the expected value of the error term given independent variable x is zero.
  • Homoscedasticity is when the variance of the error term is constant for each independent variable.

Heteroscedasticity and its Consequences

  • Heteroscedasticity refers to a scenario where the variance of the error term differs with the independent variables.

  • Under heteroscedasticity, OLS estimators are still unbiased and consistent, but the variance formulas for the OLS estimators are not valid.

  • The t-tests and F-tests are not valid under heteroscedasticity, and the OLS estimator is not the best linear unbiased estimator (BLUE).

  • Hypothesis testing for heteroscedasticity involves testing whether the expected value of the error term varies with the independent variables.

  • The Breusch-Pagan test, White test, and Alternative White test are commonly used tests for heteroscedasticity.

  • Robust standard errors should be used when heteroscedasticity is found.

  • Weighted Least Squares (WLS) can be used to estimate the model if the heteroskedasticity form is known.

  • Feasible Generalized Least Squares (FGLS) transforms the variables to get homoscedasticity if the heteroscedasticity form is not known.

  • The R-squared for the regressions of squared residuals on independent variables is used to calculate the test statistics for heteroscedasticity tests.

  • The F-test and LM-test are commonly used tests for overall significance of heteroscedasticity.

  • The regression model for price needs correction for heteroscedasticity.

  • The R-squared for the regressions of squared residuals on independent variables is used to calculate the test statistics for heteroscedasticity tests for log price.Heteroscedasticity and Regression Models

  • Heteroscedasticity is a condition where the variance of the errors is not constant across the range of values of the independent variable.

  • Heteroscedasticity can lead to biased coefficients, incorrect variance for the coefficients, and invalid t-tests and F-tests.

  • Three tests are used to identify heteroscedasticity: Breusch-Pagan test, White test, and Alternative White test.

  • The Breusch-Pagan test and White test are based on regressing the squared residuals on the independent variables.

  • The Alternative White test is based on regressing the squared residuals on the fitted values and squared fitted values.

  • If the heteroscedasticity form is known, weighted least squares (WLS) can be used to correct for it.

  • WLS assigns weights to each observation based on the inverse of the variance of the error term.

  • If the heteroscedasticity form is unknown, feasible generalized least squares (FGLS) can be used.

  • FGLS assigns weights to each observation based on the inverse of the estimated variance of the error term.

  • The coefficients are the same for ordinary least squares (OLS) and OLS with robust standard errors, but the standard errors and significance can differ.

  • The coefficients are different for OLS as compared to WLS and FGLS because of the use of weights.

  • After correcting for heteroscedasticity, the results of the various regression models (OLS, OLS with robust standard errors, WLS, and FGLS) are similar, except for the loss of significance of one coefficient in some models.

Test your understanding of linear regression with this quiz! Learn about the basics of linear regression, including the differences between simple and multiple regression models, how to calculate predicted values using the estimated regression equation, and the importance of the coefficient of determination. This quiz will also cover the least squares method, t-tests, F-tests, and the ANOVA table. Sharpen your skills and see how well you understand this fundamental statistical tool.

Make Your Own Quizzes and Flashcards

Convert your notes into interactive study material.

Get started for free
Use Quizgecko on...
Browser
Browser