Podcast
Questions and Answers
What is the consequence of omitting an important variable from a regression analysis?
What is the consequence of omitting an important variable from a regression analysis?
- The coefficient estimates will remain unbiased.
- The standard errors will become more accurate.
- The estimated coefficients will be biased and inconsistent. (correct)
- All other variable coefficients will be consistent.
What is the main assumption behind the parameter stability tests?
What is the main assumption behind the parameter stability tests?
- The model must include at least three independent variables.
- Data must be collected over multiple years.
- Parameters are constant for the entire sample period. (correct)
- Coefficient estimates are not affected by sample size.
In the Chow test, what is used to form the F-test?
In the Chow test, what is used to form the F-test?
- Difference between the sum of squared residuals (RSS) of the regressions. (correct)
- Difference between the means of the sub-periods.
- The estimated variance of the error term.
- The total number of observations in the sample.
What happens if an irrelevant variable is included in a regression model?
What happens if an irrelevant variable is included in a regression model?
When creating a dummy variable, what is the purpose of setting it to zero otherwise?
When creating a dummy variable, what is the purpose of setting it to zero otherwise?
What is the null hypothesis in the Goldfeld-Quandt test?
What is the null hypothesis in the Goldfeld-Quandt test?
When conducting the GQ test, what is the next step after splitting the sample into two sub-samples?
When conducting the GQ test, what is the next step after splitting the sample into two sub-samples?
What is the formula for the GQ test statistic?
What is the formula for the GQ test statistic?
In White's Test, what is the purpose of running the auxiliary regression?
In White's Test, what is the purpose of running the auxiliary regression?
What distribution does the test statistic from the GQ test follow under the null hypothesis?
What distribution does the test statistic from the GQ test follow under the null hypothesis?
Why might the choice of where to split the sample in the GQ test be problematic?
Why might the choice of where to split the sample in the GQ test be problematic?
How is the chi-squared statistic calculated in White’s test after running the auxiliary regression?
How is the chi-squared statistic calculated in White’s test after running the auxiliary regression?
What is indicated by the null hypothesis in the Breusch-Godfrey Test?
What is indicated by the null hypothesis in the Breusch-Godfrey Test?
What is the consequence of ignoring autocorrelation in a regression model?
What is the consequence of ignoring autocorrelation in a regression model?
Which statement is true regarding the method to correct for autocorrelation when its form is known?
Which statement is true regarding the method to correct for autocorrelation when its form is known?
What is a key characteristic of perfect multicollinearity?
What is a key characteristic of perfect multicollinearity?
In the analysis of autocorrelation, what is the significance of the test statistic exceeding the critical value?
In the analysis of autocorrelation, what is the significance of the test statistic exceeding the critical value?
What does it mean when R2 is inflated due to positively correlated residuals?
What does it mean when R2 is inflated due to positively correlated residuals?
What is a potential problem if near multicollinearity is present but ignored?
What is a potential problem if near multicollinearity is present but ignored?
Which analysis method can be used when the form of autocorrelation is unknown?
Which analysis method can be used when the form of autocorrelation is unknown?
What is the outcome if a regression model is estimated under conditions of perfect multicollinearity?
What is the outcome if a regression model is estimated under conditions of perfect multicollinearity?
What is a characteristic of regression analysis when multicollinearity is present?
What is a characteristic of regression analysis when multicollinearity is present?
Which method is NOT commonly used to measure multicollinearity?
Which method is NOT commonly used to measure multicollinearity?
What is one suggested solution to address multicollinearity?
What is one suggested solution to address multicollinearity?
What is a potential solution if the true model is a non-linear model?
What is a potential solution if the true model is a non-linear model?
Which statistical test can be used to check for functional form mis-specification in a regression model?
Which statistical test can be used to check for functional form mis-specification in a regression model?
What happens if the value of the test statistic in Ramsey’s RESET test exceeds the critical value?
What happens if the value of the test statistic in Ramsey’s RESET test exceeds the critical value?
What do skewness and kurtosis measure in a distribution?
What do skewness and kurtosis measure in a distribution?
What is a common misconception about high correlation between one of the independent variables and the dependent variable?
What is a common misconception about high correlation between one of the independent variables and the dependent variable?
Which test formalizes checking the normality of residuals?
Which test formalizes checking the normality of residuals?
What is the coefficient of kurtosis for a normal distribution?
What is the coefficient of kurtosis for a normal distribution?
Which of the following is likely a drawback of traditional solutions for multicollinearity?
Which of the following is likely a drawback of traditional solutions for multicollinearity?
What does the Bera Jarque test statistic W need to be transformed into?
What does the Bera Jarque test statistic W need to be transformed into?
What is the purpose of including higher order terms in the auxiliary regression of Ramsey's RESET test?
What is the purpose of including higher order terms in the auxiliary regression of Ramsey's RESET test?
When residuals exhibit non-normality, what is a common course of action?
When residuals exhibit non-normality, what is a common course of action?
When transforming highly correlated variables into ratios, what is the intended outcome?
When transforming highly correlated variables into ratios, what is the intended outcome?
What is one consequence of multicollinearity that affects statistical tests?
What is one consequence of multicollinearity that affects statistical tests?
What indicates the rejection of the normality assumption in residuals?
What indicates the rejection of the normality assumption in residuals?
In the context of hypothesis testing, why is normality assumed?
In the context of hypothesis testing, why is normality assumed?
What do the coefficients of skewness and kurtosis indicate when they are jointly tested for normality?
What do the coefficients of skewness and kurtosis indicate when they are jointly tested for normality?
What is the commonly used method to test for departures from normality?
What is the commonly used method to test for departures from normality?
Flashcards
Heteroscedasticity Test
Heteroscedasticity Test
A statistical test used to check if the variance of the error terms in a regression model is constant across different levels of the independent variables. It tests whether the spread of the data points around the regression line is the same across all values of the independent variable.
Goldfeld-Quandt (GQ) Test
Goldfeld-Quandt (GQ) Test
A statistical test used to detect heteroscedasticity in a regression model. It involves splitting the sample data into two groups and comparing the variances of the residuals calculated from each group.
White's Test
White's Test
A statistical test used to detect heteroscedasticity in a regression model. This test is more adaptable to different forms of heteroscedasticity and doesn't rely on specific assumptions about the pattern of heteroscedasticity.
Residuals
Residuals
Signup and view all the flashcards
Heteroscedasticity
Heteroscedasticity
Signup and view all the flashcards
Homoscedasticity
Homoscedasticity
Signup and view all the flashcards
Error Term Variance
Error Term Variance
Signup and view all the flashcards
Breusch-Godfrey Test
Breusch-Godfrey Test
Signup and view all the flashcards
Multicollinearity
Multicollinearity
Signup and view all the flashcards
Perfect Multicollinearity
Perfect Multicollinearity
Signup and view all the flashcards
Near Multicollinearity
Near Multicollinearity
Signup and view all the flashcards
Autocorrelation (in Regression)
Autocorrelation (in Regression)
Signup and view all the flashcards
Generalized Least Squares (GLS)
Generalized Least Squares (GLS)
Signup and view all the flashcards
Cochrane-Orcutt Method
Cochrane-Orcutt Method
Signup and view all the flashcards
Consequences of Ignoring Autocorrelation
Consequences of Ignoring Autocorrelation
Signup and view all the flashcards
Residual Autocorrelation
Residual Autocorrelation
Signup and view all the flashcards
What are dummy variables?
What are dummy variables?
Signup and view all the flashcards
Omission bias
Omission bias
Signup and view all the flashcards
Inclusion bias
Inclusion bias
Signup and view all the flashcards
Chow test
Chow test
Signup and view all the flashcards
Predictive failure tests
Predictive failure tests
Signup and view all the flashcards
Bera Jarque Normality Test
Bera Jarque Normality Test
Signup and view all the flashcards
Skewness
Skewness
Signup and view all the flashcards
Kurtosis
Kurtosis
Signup and view all the flashcards
Coefficient of Skewness (b1)
Coefficient of Skewness (b1)
Signup and view all the flashcards
Coefficient of Excess Kurtosis (b2-3)
Coefficient of Excess Kurtosis (b2-3)
Signup and view all the flashcards
Bera Jarque Test
Bera Jarque Test
Signup and view all the flashcards
Bera Jarque Test Statistic (W)
Bera Jarque Test Statistic (W)
Signup and view all the flashcards
Dummy Variables for Outliers
Dummy Variables for Outliers
Signup and view all the flashcards
Non-Parametric Methods
Non-Parametric Methods
Signup and view all the flashcards
Correlation Matrix
Correlation Matrix
Signup and view all the flashcards
Variance Inflationary Factor (VIF)
Variance Inflationary Factor (VIF)
Signup and view all the flashcards
Dropping a Variable
Dropping a Variable
Signup and view all the flashcards
Variable Ratio
Variable Ratio
Signup and view all the flashcards
Ramsey's RESET Test
Ramsey's RESET Test
Signup and view all the flashcards
Mis-specification of Functional Form
Mis-specification of Functional Form
Signup and view all the flashcards
Regression Residuals
Regression Residuals
Signup and view all the flashcards
Ramsey's RESET Test
Ramsey's RESET Test
Signup and view all the flashcards
Adding More Data
Adding More Data
Signup and view all the flashcards
Study Notes
Classical Linear Regression Model Assumptions and Diagnostics
- Classical linear regression models (CLRM) have assumptions for disturbance terms.
- These assumptions include:
- Expected value of the error term (εt) is zero (E(εt) = 0).
- Variance of the error term is constant (Var(εt) = σ2).
- Covariance between any two error terms is zero (cov(εi, εj) = 0 for i ≠j).
- The X matrix is non-stochastic or fixed in repeated samples.
- Errors are normally distributed (εt ~ N(0, σ2)).
Violations of CLRM Assumptions
- Studying violations of assumptions, including how to test for them, their causes, and consequences.
- Consequences can include incorrect coefficient estimates, inaccurate standard errors, and inappropriate test statistics.
- Solutions involve addressing violations or employing alternative techniques.
Assumption 1: E(εt) = 0
- The mean of the disturbances is assumed to be zero.
- Residuals are used to test this assumption, and their mean will always be zero if there's a constant term in the regression.
Assumption 2: Var(εt) = σ2
- Homoscedasticity - the variance of errors is constant (Var(εt) = σ2)
- Heteroscedasticity - the variance of errors varies.
- Detection includes methods like the Goldfeld-Quandt (GQ) test and White's test.
- The GQ test involves splitting the data, calculating residual variances, and forming a ratio for the test statistic following an F distribution.
- White's test uses an auxiliary regression based on squared residuals and regressors.
Consequences of Heteroscedasticity
- Using OLS with heteroscedasticity leads to unbiased coefficient estimates, but standard errors are wrong and inferences are flawed.
- The degree of bias in standard errors depends on the form of heteroscedasticity.
Dealing with Heteroscedasticity
- If the form of heteroscedasticity is known, generalized least squares (GLS) can be used.
- A simple illustration of GLS divides the regression by a variable related to the error variance .
Autocorrelation
- The CLRM assumes no pattern, or zero covariance, between errors (Cov(εi, εj) = 0).
- If errors have patterns, they're autocorrelated.
- Detecting autocorrelation (formal tests, such as Durbin-Watson, and Breusch-Godfrey test)
- Durbin-Watson Test (DW) tests for first-order autocorrelation, comparing errors with prior errors; ranges from 0 to 4.
- Breusch-Godfrey test is a more general, rth-order autocorrelation test.
Consequences of Ignoring Autocorrelation
- Coefficient estimates remain unbiased but are inefficient (not BLUE).
- Standard errors are inappropriate and often lead to incorrect inferences, such as incorrect conclusions about variable significance.
- R-squared values can be inflated in the presence of positively autocorrelated errors.
Remedies for Autocorrelation
- GLS techniques can be employed if the form of autocorrelation is known.
- Procedures like Cochrane-Orcutt are examples of GLS when autocorrelations are evident.
- Often modify the regression to fix autocorrelation if its form cannot be identified.
Multicollinearity
- High correlations between explanatory variables.
- Perfect multicollinearity renders coefficient estimation impossible.
- Near multicollinearity impacts coefficient standard errors (making them large) and sensitivity of the regression to specification changes.
- R-squared is often high but individual variables become less significant when multicollinearity is present.
Measuring Multicollinearity
- Method 1: Assessing the correlations between variables using a correlation matrix.
- Method 2: Analyzing the variance inflation factors (VIFs) to measure the effect of multicollinearity on independent variables.
Solutions to Multicollinearity
- Traditional techniques like ridge regression or principal component analysis.
- Some practitioners opt to ignore the issue if the model's validity is otherwise well-supported.
- Drop one of the collinear variables or transform the variables into ratios, or seek more data.
Incorrect Functional Form
- If the relationship between variables is not linear.
- Ramsey's RESET test can be used to identify non-linearity.
- This test adds higher powers of fitted values to an auxiliary regression to assess if the linearity assumption is valid by examining the R squared from the auxiliary regression.
Testing Normality
- Normality assumption implies errors are normally distributed.
- Bera-Jarque test is used, assessing skewness (b1) and kurtosis (b2); a normal distribution has zero skewness and a kurtosis of 3 (b2 = 3). A jointly zero result confirms theoretical normality.
- The test statistic is a function of these coefficients, and a large value suggests non-normality.
Solutions for Non-Normality
- Switch to a non-parametric method if normality tests produce rejection.
- Identify and consider transformations to handle non-normality or errors that are too extreme and use dummy variables for identified extreme errors.
Omission of an Important Variable or Inclusion of an Irrelevant Variable
- Omitting relevant variables leads to biased coefficient estimates in other variables.
- Including irrelevant variables increases the number of variables without improving analysis effectiveness.
Parameter Stability Test
- Assesses whether parameters in a model remain constant over the entire sample or just parts of the sample.
- Chow test is a common technique for analyzing parameter stability, essentially comparing restricted and unrestricted models.
- This approach performs a separate regression for the whole period, and each sub-part. An F ratio compares the restricted to unrestricted model.
- If the statistic exceeds the critical value, you reject the null hypothesis that the parameters are unchanging.
Studying That Suits You
Use AI to generate personalized quizzes and flashcards to suit your learning preferences.