SlideShare a Scribd company logo
PRESENTATION
ON
MULTICOLINEARITY
GROUP MEMBERS
KRISHNA CHALISE
ANKUR SHRESTHA
PAWAN KAWAN
SARITA MAHARJAN
RITU JOSHI
SONA SHRESTHA
DIPIKA SHRESTHA
Multicollinearity
• Multicollinearity occurs when two or more
independent variables in a regression model
are highly correlated to each other.
• Standard error of the OLS parameter estimate
will be higher if the corresponding independent
variable is more highly correlated to the other
independent variables in the model.
Cont…
Multicollinearity is a statistical phenomenon in
which two or more predictor variables in
a multiple regression model are
highly correlated, meaning that one can be
linearly predicted from the others with a non-
trivial degree of accuracy.
THE NATURE OF MULTICOLLINEARITY
Multicollinearity originally it meant the existence of a “perfect,” or
exact, linear relationship among some or all explanatory variables of a
regression model. For the k-variable regression involving explanatory
variable X1, X2, . . . , Xk (where X1 = 1 for all observations to allow for
the intercept term), an exact linear relationship is said to exist if the
following condition is satisfied:
λ1X1 + λ2X2 +· · ·+λkXk = 0 (10.1.1)
where λ1, λ2, . . . , λk are constants such that not all of them are zero
simultaneously.
Today, however, the term multicollinearity is used to include the case
where the X variables are intercorrelated but not perfectly so, as
follows:
λ1X1 + λ2X2 +· · ·+λ2Xk + vi = 0 (10.1.2)
where vi is a stochastic error term.
To see the difference between perfect and less than perfect
multicollinearity, assume, for example, that λ2 ≠ 0. Then,
(10.1.1) can be written as:
which shows how X2 is exactly linearly related to other
variables. In this situation, the coefficient of correlation
between the variable X2 and the linear combination on the
right side of (10.1.3) is bound to be unity.
Similarly, if λ2 ≠ 0, Eq. (10.1.2) can be written as:
which shows that X2 is not an exact linear combination of
other X’s because it is also determined by the stochastic
error term vi.
Perfect Multicollinearity
• Perfect multicollinearity occurs when there is a
perfect linear correlation between two or more
independent variables.
• When independent variable takes a constant
value in all observations.
Symptoms of Multicollinearity
• The symptoms of a multicollinearity problem
1. independent variable(s) considered critical
in explaining the model’s dependent
variable are not statistically significant
according to the tests
Symptoms of Multicollinearity
2. High R2, highly significant F-test, but few or
no statistically significant t tests
3. Parameter estimates drastically change
values and become statistically significant
when excluding some independent
variables from the regression
CAUSES OF MULTICOLLINEARITY
Causes of Multicollinearity
Data
collection
method
employed
Constraints
on the
model or in
the
population
being
sampled.
Statistical
model
specificati
on
An over
determine
d model
CONSEQUENCES OF MULTICOLLINEARITY
CONSEQUENCES OF MULTICOLLINEARITY
In cases of near or high multicollinearity, one is
likely to encounter the following consequences:
1. Although BLUE, the OLS estimators have large
variances and covariance, making precise
estimation difficult.
2. The confidence intervals tend to be much wider,
leading to the acceptance of the “zero null
hypothesis” (i.e., the true population coefficient is
zero) more readily.
1. Large Variances and Co variances of OLS Estimators
The variances and co variances of βˆ2 and βˆ3 are given by
It is apparent from (7.4.12) and (7.4.15) that as r23 tends toward 1,
that is, as co linearity increases, the variances of the two estimators
increase and in the limit when r23 = 1, they are infinite. It is equally
clear from (7.4.17) that as r23 increases toward 1, the covariance of
the two estimators also increases in absolute value.
2. Wider Confidence Intervals
Because of the large standard errors, the
confidence intervals for the relevant
population parameters tend to be larger.
Therefore, in cases of high multicollinearity,
the sample data may be compatible with a
diverse set of hypotheses. Hence, the
probability of accepting a false hypothesis
(i.e., type II error) increases.
Multicolinearity
3) t ratio of one or more coefficients tends to be
statistically insignificant.
• In cases of high collinearity, as the estimated
standard errors increased, the t values
decreased..
i.e. t= βˆ2 /S.E(βˆ2 )
Therefore, in such cases, one will increasingly
accept the null hypothesis that the relevant
true population value is zero
4) Even the t ratio is insignificant, R2 can be very high.
Since R2 is very high, we reject the null
hypothesis i.e H0: β1 =β2 = Βk = 0; due to the
significant relationship between the variables.
But since t ratios are small again we reject H0.
Therefore there is a contradictory conclusion
in the presence of multicollinearity.
5) The OLS estimators and their standard errors can be
sensitive to small changes in the data.
As long as multicollinearity is not perfect,
estimation of the regression coefficients is
possible but the OLS estimates and their
standard errors are very sensitive to even the
slightest change in the data. If we change
even only one observation, there is dramatic
change in the values.
DETECTION OF MULTICOLLINEARITY
Detection of Multicollinearity
• Multicollinearity cannot be tested; only the degree of
multicollinearity can be detected.
• Multicollinearity is a question of degree and not of kind.
The meaningful distinction is not between the presence
and the absence of multicollinearity, but between its
various degrees.
• Multicollinearity is a feature of the sample and not of the
population. Therefore, we do not “test for
multicollinearity” but we measure its degree in any
particular sample.
A. The Farrar Test
• Computation of F-ratio to test the location of
multicollinearity.
• Computation of t-ratio to test the pattern of
multicollinearity.
• Computation of chi-square to test the presence of
multicollinearity in a function with several
explanatory variables.
B. Bunch Analysis
a. Coefficient of determination, R2 , in the
presence of multicollinearity, R2 is high.
b. In the presence of muticollinearity in the
data, the partial correlation coefficients, r12
also high.
c. The high standard error of the parameters
shows the existence of multicollinearity.
Unsatisfactory Indicator of
Multicollinearity
• A large standard error may be due to the
various reasons not because of the presence of
linear relationships.
• A very high rxiyi only is neither sufficient but
nor necessary condition for the existence of
multicollinearity.
• R2 may be high to rxiyi and the result may be
highly imprecise and significant.
REMEDIAL MEASURES
Rule-of-Thumb Procedures
Other methods of remedying multicollinearity
Additional or new data
Transformation of variables
Dropping a variable(s) and specification bias
Combining cross-sectional and time series data
A priori information
REMEDIAL MEASURES
• Do nothing
– If you are only interested in prediction, multicollinearity
is not an issue.
– t-stats may be deflated, but still significant, hence
multicollinearity is not significant.
– The cure is often worse than the disease.
• Drop one or more of the multicollinear variables.
– This solution can introduce specification bias. WHY?
– In an effort to avoid specification bias a researcher can
introduce multicollinearity, hence it would be
appropriate to drop a variable.
• Transform the multicollinear variables.
– Form a linear combination of the multicollinear variables.
– Transform the equation into first differences or logs.
• Increase the sample size.
– The issue of micronumerosity.
– Micronumerosity is the problem of (n) not exceeding (k). The
symptoms are similar to the issue of multicollinearity (lack of
variation in the independent variables).
• A solution to each problem is to increase the sample size.
This will solve the problem of micronumerosity but not
necessarily the problem of multicollinearity.
• Estimate a linear model with a double-logged model.
– Such a model may be theoretically appealing since one is
estimating the elasticities.
– Logging the variables compresses the scales in which the variables
are measured hence reducing the problem of heteroskedasticity.
• In the end, you may be unable to resolve the problem of
heteroskedasticity. In this case, report White’s
heteroskedasticity-consistent variances and standard errors
and note the substantial efforts you made to reach this
conclusion
CONCLUSION
Conclusion
• Multicollinearity is a statistical phenomenon in
which there exists a perfect or exact relationship
between the predictor variables.
• When there is a perfect or exact relationship
between the predictor variables, it is difficult to
come up with reliable estimates of their individual
coefficients.
• The presence of multicollinearity can cause
serious problems with the estimation of β and the
interpretation.
• When multicollinearity is present in the data,
ordinary least square estimators are
imprecisely estimated.
• If goal is to understand how the various X
variables impact Y, then multicollinearity is a
big problem. Thus, it is very essential to detect
and solve the issue of multicollinearity before
estimating the parameter based on fitted
regression model.
• Detection of multicollinearity can be done by
examining the correlation matrix or by using
VIF.
• Remedial measures help to solve the problem
of multicollinearity.
Multicolinearity

More Related Content

PPT
Autocorrelation- Concept, Causes and Consequences
Shilpa Chaudhary
 
PPTX
Heteroscedasticity
Madurai Kamaraj University
 
PPTX
Multicollinearity PPT
GunjanKhandelwal13
 
PPTX
Autocorrelation
Pabitra Mishra
 
PPTX
Heteroscedasticity
Geethu Rangan
 
PPT
Auto Correlation Presentation
Irfan Hussain
 
PPTX
Dummy variables
Irfan Hussain
 
PDF
Heteroscedasticity
Muhammad Ali
 
Autocorrelation- Concept, Causes and Consequences
Shilpa Chaudhary
 
Heteroscedasticity
Madurai Kamaraj University
 
Multicollinearity PPT
GunjanKhandelwal13
 
Autocorrelation
Pabitra Mishra
 
Heteroscedasticity
Geethu Rangan
 
Auto Correlation Presentation
Irfan Hussain
 
Dummy variables
Irfan Hussain
 
Heteroscedasticity
Muhammad Ali
 

What's hot (20)

PPT
Multicollinearity
Bernard Asia
 
PPTX
Introduction to Econometrics
Almaszabeen Badekhan
 
PPTX
Chapter 07 - Autocorrelation.pptx
Farah Amir
 
PPTX
Basic concepts of_econometrics
SwapnaJahan
 
PPTX
Heteroscedasticity Remedial Measures.pptx
PatilDevendra5
 
DOCX
7 classical assumptions of ordinary least squares
Yugesh Dutt Panday
 
PPTX
Heteroscedasticity | Eonomics
Transweb Global Inc
 
PPT
Econometrics lecture 1st
Ishaq Ahmad
 
DOCX
Dummy variable
Akram Ali
 
PDF
Autocorrelation (1)
Manokamna Kochar
 
PPTX
Identification problem in simultaneous equations model
GarimaGupta229
 
PPTX
Specification Errors | Eonomics
Transweb Global Inc
 
PPTX
Ols by hiron
mrhasibecon
 
PPTX
Autocorrelation
ADITHYAT1
 
PPTX
Distributed lag model
Pawan Kawan
 
PPTX
Logit and Probit and Tobit model: Basic Introduction
Rabeesh Verma
 
PPT
Basic econometrics lectues_1
Nivedita Sharma
 
PPSX
Tobin's Portfolio demand for money
Prabha Panth
 
PDF
Econometrics notes (Introduction, Simple Linear regression, Multiple linear r...
Muhammad Ali
 
PPT
Cointegration and error correction model
Aditya KS
 
Multicollinearity
Bernard Asia
 
Introduction to Econometrics
Almaszabeen Badekhan
 
Chapter 07 - Autocorrelation.pptx
Farah Amir
 
Basic concepts of_econometrics
SwapnaJahan
 
Heteroscedasticity Remedial Measures.pptx
PatilDevendra5
 
7 classical assumptions of ordinary least squares
Yugesh Dutt Panday
 
Heteroscedasticity | Eonomics
Transweb Global Inc
 
Econometrics lecture 1st
Ishaq Ahmad
 
Dummy variable
Akram Ali
 
Autocorrelation (1)
Manokamna Kochar
 
Identification problem in simultaneous equations model
GarimaGupta229
 
Specification Errors | Eonomics
Transweb Global Inc
 
Ols by hiron
mrhasibecon
 
Autocorrelation
ADITHYAT1
 
Distributed lag model
Pawan Kawan
 
Logit and Probit and Tobit model: Basic Introduction
Rabeesh Verma
 
Basic econometrics lectues_1
Nivedita Sharma
 
Tobin's Portfolio demand for money
Prabha Panth
 
Econometrics notes (Introduction, Simple Linear regression, Multiple linear r...
Muhammad Ali
 
Cointegration and error correction model
Aditya KS
 
Ad

Similar to Multicolinearity (20)

PPT
Econometrics ch11
Baterdene Batchuluun
 
PPT
Econometrics_ch11.ppt
MewdedDelelegn
 
PPTX
Chapter8_Final.pptxkhnhkjllllllllllllllllllllllllllllllllllllllllllllllllllll...
DibyenduRoy49
 
PPTX
LEC11 (1).pptx
BokulHossain1
 
PDF
Multicollinearity econometrics semester 4 Delhi University
killerharsh4100
 
PDF
Multicolliniaritymultipleregression
Hazwani Husna
 
PPTX
Multicollinearity meaning and its Consequences
EzhilarasanG1
 
PPTX
Multicollinearity.pptx this is a presentation of hetro.
vrao95787
 
PDF
Multicollinearity1
Muhammad Ali
 
PPTX
MULTICOLLINERITY.pptx
YanYingLoh
 
PDF
Perfect and imperfect multicollinearitya) Define perfect multicol.pdf
akagencies2014
 
PPTX
All About Econometrics
Ratul Basak
 
PDF
Define Multicollinearity in the following terms a. In what .pdf
alshaikhkhanzariarts
 
PDF
Chapter5.pdf.pdf
ROBERTOENRIQUEGARCAA1
 
PPT
Econometrics chapter three and four finance.ppt
etebarkhmichale
 
PPT
Multiple Regression.ppt
EkoGaniarto
 
PPTX
Multiple regression .pptx
Victoria Bozhenko
 
PPTX
Econometrics.pptx
MahafuzurRahmanMafi
 
PPT
Econometric model ing
Matt Grant
 
PPTX
Chapter 5 (1) Multi Collinearity.pptx econometrics
ImtiazAli27143
 
Econometrics ch11
Baterdene Batchuluun
 
Econometrics_ch11.ppt
MewdedDelelegn
 
Chapter8_Final.pptxkhnhkjllllllllllllllllllllllllllllllllllllllllllllllllllll...
DibyenduRoy49
 
LEC11 (1).pptx
BokulHossain1
 
Multicollinearity econometrics semester 4 Delhi University
killerharsh4100
 
Multicolliniaritymultipleregression
Hazwani Husna
 
Multicollinearity meaning and its Consequences
EzhilarasanG1
 
Multicollinearity.pptx this is a presentation of hetro.
vrao95787
 
Multicollinearity1
Muhammad Ali
 
MULTICOLLINERITY.pptx
YanYingLoh
 
Perfect and imperfect multicollinearitya) Define perfect multicol.pdf
akagencies2014
 
All About Econometrics
Ratul Basak
 
Define Multicollinearity in the following terms a. In what .pdf
alshaikhkhanzariarts
 
Chapter5.pdf.pdf
ROBERTOENRIQUEGARCAA1
 
Econometrics chapter three and four finance.ppt
etebarkhmichale
 
Multiple Regression.ppt
EkoGaniarto
 
Multiple regression .pptx
Victoria Bozhenko
 
Econometrics.pptx
MahafuzurRahmanMafi
 
Econometric model ing
Matt Grant
 
Chapter 5 (1) Multi Collinearity.pptx econometrics
ImtiazAli27143
 
Ad

More from Pawan Kawan (20)

PPTX
Efficient market hypothesis
Pawan Kawan
 
PPTX
Profile of Robert mondavi
Pawan Kawan
 
PPTX
Fredrick w. smith
Pawan Kawan
 
PPTX
Development of financial institutions in Nepal
Pawan Kawan
 
DOCX
Econometrics
Pawan Kawan
 
DOCX
Gross domestic product of nepal
Pawan Kawan
 
DOCX
Operation management
Pawan Kawan
 
DOCX
Outsourcing
Pawan Kawan
 
PPTX
The fixed exchange rate system
Pawan Kawan
 
PPTX
Profitability of nepalese commercial bank
Pawan Kawan
 
PPTX
Operation management
Pawan Kawan
 
PPTX
Hard rock cafe
Pawan Kawan
 
PPT
Econometrics and business forecasting
Pawan Kawan
 
PPTX
Determination of interest rate
Pawan Kawan
 
PPTX
Asian financial crisis
Pawan Kawan
 
DOCX
Internship report (nrb)
Pawan Kawan
 
PPTX
Three rules for making a company truly great
Pawan Kawan
 
PPTX
Internship at nepal rastra bank ( nrb )
Pawan Kawan
 
PPTX
Creating and managing bank service
Pawan Kawan
 
PPTX
Role and function of nepal rastra bank
Pawan Kawan
 
Efficient market hypothesis
Pawan Kawan
 
Profile of Robert mondavi
Pawan Kawan
 
Fredrick w. smith
Pawan Kawan
 
Development of financial institutions in Nepal
Pawan Kawan
 
Econometrics
Pawan Kawan
 
Gross domestic product of nepal
Pawan Kawan
 
Operation management
Pawan Kawan
 
Outsourcing
Pawan Kawan
 
The fixed exchange rate system
Pawan Kawan
 
Profitability of nepalese commercial bank
Pawan Kawan
 
Operation management
Pawan Kawan
 
Hard rock cafe
Pawan Kawan
 
Econometrics and business forecasting
Pawan Kawan
 
Determination of interest rate
Pawan Kawan
 
Asian financial crisis
Pawan Kawan
 
Internship report (nrb)
Pawan Kawan
 
Three rules for making a company truly great
Pawan Kawan
 
Internship at nepal rastra bank ( nrb )
Pawan Kawan
 
Creating and managing bank service
Pawan Kawan
 
Role and function of nepal rastra bank
Pawan Kawan
 

Recently uploaded (20)

PPTX
Tips Management in Odoo 18 POS - Odoo Slides
Celine George
 
PPTX
How to Close Subscription in Odoo 18 - Odoo Slides
Celine George
 
PPTX
Command Palatte in Odoo 18.1 Spreadsheet - Odoo Slides
Celine George
 
PPTX
Care of patients with elImination deviation.pptx
AneetaSharma15
 
PPTX
HEALTH CARE DELIVERY SYSTEM - UNIT 2 - GNM 3RD YEAR.pptx
Priyanshu Anand
 
PPTX
family health care settings home visit - unit 6 - chn 1 - gnm 1st year.pptx
Priyanshu Anand
 
PPTX
Applications of matrices In Real Life_20250724_091307_0000.pptx
gehlotkrish03
 
PDF
Antianginal agents, Definition, Classification, MOA.pdf
Prerana Jadhav
 
PPTX
Cleaning Validation Ppt Pharmaceutical validation
Ms. Ashatai Patil
 
DOCX
Modul Ajar Deep Learning Bahasa Inggris Kelas 11 Terbaru 2025
wahyurestu63
 
PDF
Biological Classification Class 11th NCERT CBSE NEET.pdf
NehaRohtagi1
 
PPTX
Python-Application-in-Drug-Design by R D Jawarkar.pptx
Rahul Jawarkar
 
PPTX
CDH. pptx
AneetaSharma15
 
PPTX
Artificial-Intelligence-in-Drug-Discovery by R D Jawarkar.pptx
Rahul Jawarkar
 
PDF
BÀI TẬP TEST BỔ TRỢ THEO TỪNG CHỦ ĐỀ CỦA TỪNG UNIT KÈM BÀI TẬP NGHE - TIẾNG A...
Nguyen Thanh Tu Collection
 
PPTX
Basics and rules of probability with real-life uses
ravatkaran694
 
PDF
Review of Related Literature & Studies.pdf
Thelma Villaflores
 
PPTX
CARE OF UNCONSCIOUS PATIENTS .pptx
AneetaSharma15
 
PPTX
Continental Accounting in Odoo 18 - Odoo Slides
Celine George
 
PDF
What is CFA?? Complete Guide to the Chartered Financial Analyst Program
sp4989653
 
Tips Management in Odoo 18 POS - Odoo Slides
Celine George
 
How to Close Subscription in Odoo 18 - Odoo Slides
Celine George
 
Command Palatte in Odoo 18.1 Spreadsheet - Odoo Slides
Celine George
 
Care of patients with elImination deviation.pptx
AneetaSharma15
 
HEALTH CARE DELIVERY SYSTEM - UNIT 2 - GNM 3RD YEAR.pptx
Priyanshu Anand
 
family health care settings home visit - unit 6 - chn 1 - gnm 1st year.pptx
Priyanshu Anand
 
Applications of matrices In Real Life_20250724_091307_0000.pptx
gehlotkrish03
 
Antianginal agents, Definition, Classification, MOA.pdf
Prerana Jadhav
 
Cleaning Validation Ppt Pharmaceutical validation
Ms. Ashatai Patil
 
Modul Ajar Deep Learning Bahasa Inggris Kelas 11 Terbaru 2025
wahyurestu63
 
Biological Classification Class 11th NCERT CBSE NEET.pdf
NehaRohtagi1
 
Python-Application-in-Drug-Design by R D Jawarkar.pptx
Rahul Jawarkar
 
CDH. pptx
AneetaSharma15
 
Artificial-Intelligence-in-Drug-Discovery by R D Jawarkar.pptx
Rahul Jawarkar
 
BÀI TẬP TEST BỔ TRỢ THEO TỪNG CHỦ ĐỀ CỦA TỪNG UNIT KÈM BÀI TẬP NGHE - TIẾNG A...
Nguyen Thanh Tu Collection
 
Basics and rules of probability with real-life uses
ravatkaran694
 
Review of Related Literature & Studies.pdf
Thelma Villaflores
 
CARE OF UNCONSCIOUS PATIENTS .pptx
AneetaSharma15
 
Continental Accounting in Odoo 18 - Odoo Slides
Celine George
 
What is CFA?? Complete Guide to the Chartered Financial Analyst Program
sp4989653
 

Multicolinearity

  • 1. PRESENTATION ON MULTICOLINEARITY GROUP MEMBERS KRISHNA CHALISE ANKUR SHRESTHA PAWAN KAWAN SARITA MAHARJAN RITU JOSHI SONA SHRESTHA DIPIKA SHRESTHA
  • 2. Multicollinearity • Multicollinearity occurs when two or more independent variables in a regression model are highly correlated to each other. • Standard error of the OLS parameter estimate will be higher if the corresponding independent variable is more highly correlated to the other independent variables in the model.
  • 3. Cont… Multicollinearity is a statistical phenomenon in which two or more predictor variables in a multiple regression model are highly correlated, meaning that one can be linearly predicted from the others with a non- trivial degree of accuracy.
  • 4. THE NATURE OF MULTICOLLINEARITY Multicollinearity originally it meant the existence of a “perfect,” or exact, linear relationship among some or all explanatory variables of a regression model. For the k-variable regression involving explanatory variable X1, X2, . . . , Xk (where X1 = 1 for all observations to allow for the intercept term), an exact linear relationship is said to exist if the following condition is satisfied: λ1X1 + λ2X2 +· · ·+λkXk = 0 (10.1.1) where λ1, λ2, . . . , λk are constants such that not all of them are zero simultaneously. Today, however, the term multicollinearity is used to include the case where the X variables are intercorrelated but not perfectly so, as follows: λ1X1 + λ2X2 +· · ·+λ2Xk + vi = 0 (10.1.2) where vi is a stochastic error term.
  • 5. To see the difference between perfect and less than perfect multicollinearity, assume, for example, that λ2 ≠ 0. Then, (10.1.1) can be written as: which shows how X2 is exactly linearly related to other variables. In this situation, the coefficient of correlation between the variable X2 and the linear combination on the right side of (10.1.3) is bound to be unity. Similarly, if λ2 ≠ 0, Eq. (10.1.2) can be written as: which shows that X2 is not an exact linear combination of other X’s because it is also determined by the stochastic error term vi.
  • 6. Perfect Multicollinearity • Perfect multicollinearity occurs when there is a perfect linear correlation between two or more independent variables. • When independent variable takes a constant value in all observations.
  • 7. Symptoms of Multicollinearity • The symptoms of a multicollinearity problem 1. independent variable(s) considered critical in explaining the model’s dependent variable are not statistically significant according to the tests
  • 8. Symptoms of Multicollinearity 2. High R2, highly significant F-test, but few or no statistically significant t tests 3. Parameter estimates drastically change values and become statistically significant when excluding some independent variables from the regression
  • 10. Causes of Multicollinearity Data collection method employed Constraints on the model or in the population being sampled. Statistical model specificati on An over determine d model
  • 12. CONSEQUENCES OF MULTICOLLINEARITY In cases of near or high multicollinearity, one is likely to encounter the following consequences: 1. Although BLUE, the OLS estimators have large variances and covariance, making precise estimation difficult. 2. The confidence intervals tend to be much wider, leading to the acceptance of the “zero null hypothesis” (i.e., the true population coefficient is zero) more readily.
  • 13. 1. Large Variances and Co variances of OLS Estimators The variances and co variances of βˆ2 and βˆ3 are given by It is apparent from (7.4.12) and (7.4.15) that as r23 tends toward 1, that is, as co linearity increases, the variances of the two estimators increase and in the limit when r23 = 1, they are infinite. It is equally clear from (7.4.17) that as r23 increases toward 1, the covariance of the two estimators also increases in absolute value.
  • 14. 2. Wider Confidence Intervals Because of the large standard errors, the confidence intervals for the relevant population parameters tend to be larger. Therefore, in cases of high multicollinearity, the sample data may be compatible with a diverse set of hypotheses. Hence, the probability of accepting a false hypothesis (i.e., type II error) increases.
  • 16. 3) t ratio of one or more coefficients tends to be statistically insignificant. • In cases of high collinearity, as the estimated standard errors increased, the t values decreased.. i.e. t= βˆ2 /S.E(βˆ2 ) Therefore, in such cases, one will increasingly accept the null hypothesis that the relevant true population value is zero
  • 17. 4) Even the t ratio is insignificant, R2 can be very high. Since R2 is very high, we reject the null hypothesis i.e H0: β1 =β2 = Βk = 0; due to the significant relationship between the variables. But since t ratios are small again we reject H0. Therefore there is a contradictory conclusion in the presence of multicollinearity.
  • 18. 5) The OLS estimators and their standard errors can be sensitive to small changes in the data. As long as multicollinearity is not perfect, estimation of the regression coefficients is possible but the OLS estimates and their standard errors are very sensitive to even the slightest change in the data. If we change even only one observation, there is dramatic change in the values.
  • 20. Detection of Multicollinearity • Multicollinearity cannot be tested; only the degree of multicollinearity can be detected. • Multicollinearity is a question of degree and not of kind. The meaningful distinction is not between the presence and the absence of multicollinearity, but between its various degrees. • Multicollinearity is a feature of the sample and not of the population. Therefore, we do not “test for multicollinearity” but we measure its degree in any particular sample.
  • 21. A. The Farrar Test • Computation of F-ratio to test the location of multicollinearity. • Computation of t-ratio to test the pattern of multicollinearity. • Computation of chi-square to test the presence of multicollinearity in a function with several explanatory variables.
  • 22. B. Bunch Analysis a. Coefficient of determination, R2 , in the presence of multicollinearity, R2 is high. b. In the presence of muticollinearity in the data, the partial correlation coefficients, r12 also high. c. The high standard error of the parameters shows the existence of multicollinearity.
  • 23. Unsatisfactory Indicator of Multicollinearity • A large standard error may be due to the various reasons not because of the presence of linear relationships. • A very high rxiyi only is neither sufficient but nor necessary condition for the existence of multicollinearity. • R2 may be high to rxiyi and the result may be highly imprecise and significant.
  • 25. Rule-of-Thumb Procedures Other methods of remedying multicollinearity Additional or new data Transformation of variables Dropping a variable(s) and specification bias Combining cross-sectional and time series data A priori information
  • 26. REMEDIAL MEASURES • Do nothing – If you are only interested in prediction, multicollinearity is not an issue. – t-stats may be deflated, but still significant, hence multicollinearity is not significant. – The cure is often worse than the disease. • Drop one or more of the multicollinear variables. – This solution can introduce specification bias. WHY? – In an effort to avoid specification bias a researcher can introduce multicollinearity, hence it would be appropriate to drop a variable.
  • 27. • Transform the multicollinear variables. – Form a linear combination of the multicollinear variables. – Transform the equation into first differences or logs. • Increase the sample size. – The issue of micronumerosity. – Micronumerosity is the problem of (n) not exceeding (k). The symptoms are similar to the issue of multicollinearity (lack of variation in the independent variables). • A solution to each problem is to increase the sample size. This will solve the problem of micronumerosity but not necessarily the problem of multicollinearity.
  • 28. • Estimate a linear model with a double-logged model. – Such a model may be theoretically appealing since one is estimating the elasticities. – Logging the variables compresses the scales in which the variables are measured hence reducing the problem of heteroskedasticity. • In the end, you may be unable to resolve the problem of heteroskedasticity. In this case, report White’s heteroskedasticity-consistent variances and standard errors and note the substantial efforts you made to reach this conclusion
  • 30. Conclusion • Multicollinearity is a statistical phenomenon in which there exists a perfect or exact relationship between the predictor variables. • When there is a perfect or exact relationship between the predictor variables, it is difficult to come up with reliable estimates of their individual coefficients. • The presence of multicollinearity can cause serious problems with the estimation of β and the interpretation.
  • 31. • When multicollinearity is present in the data, ordinary least square estimators are imprecisely estimated. • If goal is to understand how the various X variables impact Y, then multicollinearity is a big problem. Thus, it is very essential to detect and solve the issue of multicollinearity before estimating the parameter based on fitted regression model.
  • 32. • Detection of multicollinearity can be done by examining the correlation matrix or by using VIF. • Remedial measures help to solve the problem of multicollinearity.