The trick here is "unique variance", and
what you need to read up on is "suppressor variables".
If you look for it, you will see that one of these variables
has a partial-correlation that is in the opposite direction
from its 0-order correlation. This happens whenever
the *difference* (or ratio) of the two correlated predictors
is also predictive. - An easy clue that this is happening is
that a standardized beta is greater than 1.0.
The best solution that I know of is to create a new variable,
a composite score that is some (weighted) difference or ratio
(or log-ratio) of the two correlated variables -- if there is
such a composite that makes particularly good sense for the
variables, use that version.
--
Rich Ulrich
Date: Fri, 27 Jul 2012 10:06:15 -0700
From:
[hidden email]Subject: Part correlations and multicollinearity?
To:
[hidden email]
Hello to all,
I have a regression model where two variables are highly correlated, at ~ .8, and the Tolerance is low .2 - .3 and VIF high, ~4-5.
Does it mean that the variables don’t have such a large independent contribution to explaining the outcome? Even if there seems to be multicollinearity for the two variables, these same two variables have large part and semipartial correlation coefficients, the largest amongst all 10 variables included (5 controls).
Can I say that despite the multicollinearity, each variable explains a unique proportion in the outcome only based on these part correlations? I have read somewhere that <The more “tolerant” a variable is (i.e. the less highly correlated it is with the other IVs), the greater its unique contribution to R2 will be>, but the case here is reverse, these variables have low tolerance but high contribution to explaining the criterion.
Is this interpretation correct?
Thank you in advance.
|