Podcast
Questions and Answers
What is the distribution of the LM statistic?
What is the distribution of the LM statistic?
- Uniform distribution
- $LM ext{ ∼ } χ^2q$ (correct)
- $ar{LM} ext{ ∼ }rac{ ext{1}}{ χ^2q }$
- Normal distribution
Under which conditions is OLS considered asymptotically efficient?
Under which conditions is OLS considered asymptotically efficient?
- When estimators are inconsistent
- In the presence of heteroskedasticity
- Under the Gauss-Markov assumptions (correct)
- When errors are normally distributed
What should one expect regarding the results from an F test and an LM test with a large sample?
What should one expect regarding the results from an F test and an LM test with a large sample?
- The F test will always be more powerful
- The LM test is irrelevant in large samples
- They should be similar (correct)
- They will be identical
What happens to the OLS efficiency conclusion if the error term is not homoskedastic?
What happens to the OLS efficiency conclusion if the error term is not homoskedastic?
What is used to choose a critical value in the LM test?
What is used to choose a critical value in the LM test?
What does the Central Limit Theorem imply about OLS estimators as the sample size increases?
What does the Central Limit Theorem imply about OLS estimators as the sample size increases?
According to asymptotic normality, what is the distribution of the standardized OLS estimator $Z$?
According to asymptotic normality, what is the distribution of the standardized OLS estimator $Z$?
Which assumptions must hold for the OLS estimators to be asymptotically normal under the Gauss-Markov assumptions?
Which assumptions must hold for the OLS estimators to be asymptotically normal under the Gauss-Markov assumptions?
In the context of asymptotic inference, what does $P(Z < z) o ext{Φ}(z)$ represent?
In the context of asymptotic inference, what does $P(Z < z) o ext{Φ}(z)$ represent?
What does $σ̂^2$ represent in the context of asymptotic normality?
What does $σ̂^2$ represent in the context of asymptotic normality?
What does the asymptotic standard error for coefficient estimates depend on?
What does the asymptotic standard error for coefficient estimates depend on?
In the context of the LM statistic, what is the purpose of the auxiliary regression?
In the context of the LM statistic, what is the purpose of the auxiliary regression?
What does the Lagrange Multiplier (LM) statistic primarily rely on for inference?
What does the Lagrange Multiplier (LM) statistic primarily rely on for inference?
What does the model's null hypothesis generally state when testing with the LM statistic?
What does the model's null hypothesis generally state when testing with the LM statistic?
What effect does sample size (N) have on the standard errors in large sample theory?
What effect does sample size (N) have on the standard errors in large sample theory?
What does it mean for an estimator to be consistent?
What does it mean for an estimator to be consistent?
Under which conditions is the OLS estimator considered to be BLUE?
Under which conditions is the OLS estimator considered to be BLUE?
What is the implication of having a variance tending to zero for an estimator?
What is the implication of having a variance tending to zero for an estimator?
What is the purpose of taking the probability limit (plim) in establishing consistency?
What is the purpose of taking the probability limit (plim) in establishing consistency?
What assumption regarding the matrix of independent variables X is made to establish the consistency of the OLS estimator?
What assumption regarding the matrix of independent variables X is made to establish the consistency of the OLS estimator?
In the context of the SLR model, how is the OLS estimator expressed?
In the context of the SLR model, how is the OLS estimator expressed?
What does the term 'asymptotic inference' refer to in statistical estimation?
What does the term 'asymptotic inference' refer to in statistical estimation?
Which of the Gauss-Markov assumptions is critical for both unbiasedness and consistency of OLS estimators?
Which of the Gauss-Markov assumptions is critical for both unbiasedness and consistency of OLS estimators?
What does the term plimw = 0 signify in the context of the document?
What does the term plimw = 0 signify in the context of the document?
Which of the following is NOT a condition required for unbiasedness in OLS?
Which of the following is NOT a condition required for unbiasedness in OLS?
What does inconsistency in OLS imply in terms of sample size?
What does inconsistency in OLS imply in terms of sample size?
What implication does the assumption of normality of errors have on sampling distributions?
What implication does the assumption of normality of errors have on sampling distributions?
Which statement about OLS being BLUE is correct?
Which statement about OLS being BLUE is correct?
Which of the following must be assumed about the variances for consistency in OLS?
Which of the following must be assumed about the variances for consistency in OLS?
What characteristic of a normally distributed error term is significant in a regression model?
What characteristic of a normally distributed error term is significant in a regression model?
What conclusion can be drawn if the distribution of the dependent variable is skewed?
What conclusion can be drawn if the distribution of the dependent variable is skewed?
What happens to the sample moments as N increases?
What happens to the sample moments as N increases?
What does the equation plim$etâ$ = plim$eta$ + plim$rac{1}{N}X' u$ indicate about the OLS estimator?
What does the equation plim$etâ$ = plim$eta$ + plim$rac{1}{N}X' u$ indicate about the OLS estimator?
What condition must hold for Cov($x_1$, $u$) to ensure consistent estimation?
What condition must hold for Cov($x_1$, $u$) to ensure consistent estimation?
In the context of the MLR model, what does $etâ = (X'X)^{-1}X'y$ denote?
In the context of the MLR model, what does $etâ = (X'X)^{-1}X'y$ denote?
What is the implication of plim$rac{1}{N}X' u = 0$?
What is the implication of plim$rac{1}{N}X' u = 0$?
What does the term $Var(w) = E(Var(w|X)) + Var[E(w|X)]$ represent?
What does the term $Var(w) = E(Var(w|X)) + Var[E(w|X)]$ represent?
What is the role of OLS in estimating parameters?
What is the role of OLS in estimating parameters?
Which mathematical concept is involved in the OLS estimator's calculation?
Which mathematical concept is involved in the OLS estimator's calculation?
Which of the following statements about variance in the context of the OLS model is true?
Which of the following statements about variance in the context of the OLS model is true?
What does the notation $E[E(w|X)] = 0$ imply about the expected value of the error term?
What does the notation $E[E(w|X)] = 0$ imply about the expected value of the error term?
Flashcards
Consistency of OLS Estimator
Consistency of OLS Estimator
Under the Gauss-Markov assumptions (MLR.1-MLR.5), the OLS estimator (β̂j) is consistent for the true parameter (βj) for all explanatory variables (j from 1 to K).
Consistency
Consistency
As the sample size (N) gets larger, the distribution of an estimator converges to the true parameter value. This means the estimator's average (mean) approaches the true parameter, and its variability diminishes.
Gauss-Markov assumptions
Gauss-Markov assumptions
A set of assumptions (MLR 1-5) necessary for OLS to be both unbiased and efficient (BLUE) in linear regression models.
Probability Limit (plim)
Probability Limit (plim)
Signup and view all the flashcards
Law of Large Numbers
Law of Large Numbers
Signup and view all the flashcards
OLS estimator (β̂)
OLS estimator (β̂)
Signup and view all the flashcards
BLUE
BLUE
Signup and view all the flashcards
SLR model
SLR model
Signup and view all the flashcards
Consistency of OLS
Consistency of OLS
Signup and view all the flashcards
plim β̂1
plim β̂1
Signup and view all the flashcards
plim β1
plim β1
Signup and view all the flashcards
β̂ = (X'X)^-1 X'y
β̂ = (X'X)^-1 X'y
Signup and view all the flashcards
MLR model (Multiple Linear Regression)
MLR model (Multiple Linear Regression)
Signup and view all the flashcards
Cov(x1, u) = 0
Cov(x1, u) = 0
Signup and view all the flashcards
Var(x1)
Var(x1)
Signup and view all the flashcards
E(ui|xi) = 0
E(ui|xi) = 0
Signup and view all the flashcards
N
N
Signup and view all the flashcards
plim
plim
Signup and view all the flashcards
LM Statistic
LM Statistic
Signup and view all the flashcards
LM Statistic Distribution
LM Statistic Distribution
Signup and view all the flashcards
F Test vs. LM Test
F Test vs. LM Test
Signup and view all the flashcards
Asymptotic Efficiency
Asymptotic Efficiency
Signup and view all the flashcards
Gauss-Markov Assumptions
Gauss-Markov Assumptions
Signup and view all the flashcards
Asymptotic Normality of OLS
Asymptotic Normality of OLS
Signup and view all the flashcards
Central Limit Theorem (CLT)
Central Limit Theorem (CLT)
Signup and view all the flashcards
Asymptotic Standard Error
Asymptotic Standard Error
Signup and view all the flashcards
OLS Estimator (β̂)
OLS Estimator (β̂)
Signup and view all the flashcards
Consistent Estimator
Consistent Estimator
Signup and view all the flashcards
Consistency
Consistency
Signup and view all the flashcards
Zero Conditional Mean
Zero Conditional Mean
Signup and view all the flashcards
Zero Mean
Zero Mean
Signup and view all the flashcards
Zero Correlation
Zero Correlation
Signup and view all the flashcards
Asymptotic Inference
Asymptotic Inference
Signup and view all the flashcards
Normality Assumption
Normality Assumption
Signup and view all the flashcards
Inconsistent Estimator
Inconsistent Estimator
Signup and view all the flashcards
Probability Limit
Probability Limit
Signup and view all the flashcards
Asymptotic Standard Error
Asymptotic Standard Error
Signup and view all the flashcards
LM Statistic
LM Statistic
Signup and view all the flashcards
LM Statistic (Auxiliary Regression)
LM Statistic (Auxiliary Regression)
Signup and view all the flashcards
Restricted Model
Restricted Model
Signup and view all the flashcards
Null Hypothesis (H0)
Null Hypothesis (H0)
Signup and view all the flashcards
Study Notes
Unit 6: Asymptotic Theory and Properties
- The unit covers large sample properties and asymptotic inference in econometrics.
- Large sample properties include consistency.
- Asymptotic inference is based on the Central Limit Theorem.
Large Sample Properties: Consistency
- Ordinary Least Squares (OLS) is Best Linear Unbiased Estimator (BLUE) under Gauss-Markov assumptions, but not always in other cases.
- In other cases, consistent estimators might be used.
- Consistent estimators mean that, as the sample size (N) approaches infinity, the estimator's distribution collapses to the parameter value.
- Consequently, the mean of the estimator approaches the parameter value, and the variance tends to zero as N tends to infinity.
Consistency of the OLS Estimator
- Under Gauss-Markov assumptions (MLR.1-MLR.5), the OLS estimator is consistent for each parameter.
- The same assumptions ensuring unbiasedness also imply consistency.
- The probability limit (plim) is used to establish consistency.
- The second moments of the independent variables (X) must be finite.
Proving Consistency – The SLR Model
- The OLS estimator can be written using sample moments.
- Applying the law of large numbers, sample moments converge in probability to population counterparts as N increases.
- This shows the OLS estimator converges to the true parameter value.
Proving Consistency – The MLR Model
- The OLS estimator is expressed using matrix notation.
- Applying the law of large numbers, the estimator converges in probability to the true parameter value.
Convergence – The Full Proof
- The OLS estimator is presented in terms of sample moments and the population parameter.
- The variance calculation demonstrates how the estimator's variance goes to zero as the sample size increases.
- In conclusion, the estimator converges in mean square to zero, resulting in the probability limit (plim) of the estimator being the true parameter value.
A Weaker Assumption
- For unbiasedness, a zero conditional mean (E(u|X) = 0) of the error term was assumed.
- Consistency only requires a zero mean and zero correlation between independent variables and the error (E(u)= 0 and Cov(Xj, u) = 0 ).
- Without these assumptions, OLS is inconsistent.
Asymptotic Inference
- Under Classical Linear Model (CLM) assumptions, sampling distributions are normally distributed. This allows the derivation of t and F distributions for testing.
- Normality assumption implies normal distribution of y given x's.
Asymptotic Inference (Continued)
- Clearly skewed variables (e.g., wages, arrests) cannot be normally distributed since the normal distribution is symmetric.
- Normality is not necessary for OLS to be BLUE, only for inference (statistical significance).
- Using Central Limit Theorem, OLS estimator is asymptotically normally distributed.
Asymptotic Normality - I
- Under Gauss-Markov assumptions (MLR.1-MLR.5), the OLS estimators have asymptotic normal distributions.
- The asymptotic variance of 𝛽j is σ2/aj, where σ2 is the error variance and aj is the plim(1/(X'X))j,i.
- ô is a consistent estimator of σ2.
Asymptotic Normality - II (More Generally)
- Asymptotic normality holds with independent, identically distributed (iid) errors and finite variance.
- Other theorems (law of large numbers for plims and Central Limit Theorem for asymptotic normality), are used to derive the asymptotic results.
Asymptotic Normality - III
- Since the t-distribution approximates normal distribution with large degrees of freedom, t-tests can be used asymptotically.
- Homoskedasticity is still required for the asymptotic t-test, despite normality not being required.
Asymptotic Standard Errors
- If the error term is not normally distributed, the standard error can be referred to as an asymptotic standard error.
- The formula for the asymptotic standard error shows it shrinks proportionally to the inverse of the square root of the sample size (√N).
Lagrange Multiplier Statistic - I and II and III
- The Lagrange multiplier (LM) statistic provides an alternative to the F-statistic for testing multiple exclusion restrictions.
- It's based on an "auxiliary regression" and is sometimes called an NR² statistic.
- The LM statistic can be calculated by running a restricted model, finding residuals from that model, and then regressing these residuals on all independent variables in the original model.
- Results are similar to F-tests for large samples, however, they are not identical for testing a single exclusion restriction.
Asymptotic Efficiency
- Other estimators may be consistent; however, OLS has smallest asymptotic variances under Gauss-Markov assumptions.
- The conclusion of OLS' asymptotic efficiency depends on the assumptions, particularly homoskedasticity.
Studying That Suits You
Use AI to generate personalized quizzes and flashcards to suit your learning preferences.