sriram Narayanan
2006-01-21 22:30:45 UTC
Hi All:
I have been seeing the archives for a discussion of multicollinearity.
I think I want to divide this into two parts.
1. Diagnose if there is a multicollinearity problem
2. Address the issue if multicollinearity is an issue.
I am dividing this because I want to bring some clarity into my head
about this issue, had some questions on them. I probably may be saying
something very silly but I just want to learn more about this problem
because I am facing such issues with my model that I have been
thinking. My model contains an X term and an X^2 term. (.97
correlation)
On diagnosis: There are two ways I know till now:
Method 1) Some earlier messages (I cant trace this message again
except I read it otherwise I would acknowledge the source) seemed to
suggest that one should look at the level of correlation between the
estimators. In lisrel I think PC option gives this. Very high
estimator correlations indicate that there is a problem with
multicollinearity. For example my model with AMOS converges with a
very bad fit, and I have problems with correlations between the
estimates involving the paths from x to the latent variable and x^2 to
the latent variable. However my model in LISREL converges fine.
Method 2) If the completely standardized coefficient is greater than
1 it can indicate multicollinearity as in one of Prof Joreskog's
article.
http://www.ssicentral.com/lisrel/techdocs/HowLargeCanaStandardizedCoefficientbe.pdf
Even though the correlations are high my completely standardized
correlations are less than 1.
Questions:
What happens if one has a completely standardized coefficient of <1
but high correlations between the estimates. ?
Having a completely standardized coefficient of grater than 1 could be
a strong indicator of multicollinearity but can one also say that
having a less than 1 guarantees that there is no multicollinearity?
What is the theory behind the reasons for high estimator correlations
leading to multicollinearity. I will appreciate if there are any
references.
In my model diagnostic 2 seems fine even if the two indicators are
highly correlated in my case (the correlation between them is .97).
however the estimates are strongly correlated *,95 upwards)
Would I think that the collinearity problem is not serious?
On correcting multicollinearity:
Ed's suggested that one can do centering as has been suggested by
authors like Chronbach (1987, Psychological Bulletin) in this
listserv.
I have tried that strategy and it seemed to me that somehow in my data
it does not work I have a .96 correlation before centering and a .76
after centering. It does go down but I am not sure if this is low
enough. When I use this method diagnostic 1 still has some problems
while disgnstic 2 is fine in my model.
The other way is standardization that I found in an article by
Dunlap, William P. AND Edward R. Kemery (1988), "Effects of Predictor ...
INTERACTIONS AND Moderator Effects," Psychological Bulletin, 114 (2), 376-390
I have done log transformations to get my variable to be normal. If I
mean center this would it not affect my interpretation ?
Some clarity on this will be highly welcome. I will be very grateful.
Thanks
Sriram
--------------------------------------------------------------
To unsubscribe from SEMNET, send email to ***@bama.ua.edu
with the body of the message as: SIGNOFF SEMNET
Search the archives at http://bama.ua.edu/archives/semnet.html
I have been seeing the archives for a discussion of multicollinearity.
I think I want to divide this into two parts.
1. Diagnose if there is a multicollinearity problem
2. Address the issue if multicollinearity is an issue.
I am dividing this because I want to bring some clarity into my head
about this issue, had some questions on them. I probably may be saying
something very silly but I just want to learn more about this problem
because I am facing such issues with my model that I have been
thinking. My model contains an X term and an X^2 term. (.97
correlation)
On diagnosis: There are two ways I know till now:
Method 1) Some earlier messages (I cant trace this message again
except I read it otherwise I would acknowledge the source) seemed to
suggest that one should look at the level of correlation between the
estimators. In lisrel I think PC option gives this. Very high
estimator correlations indicate that there is a problem with
multicollinearity. For example my model with AMOS converges with a
very bad fit, and I have problems with correlations between the
estimates involving the paths from x to the latent variable and x^2 to
the latent variable. However my model in LISREL converges fine.
Method 2) If the completely standardized coefficient is greater than
1 it can indicate multicollinearity as in one of Prof Joreskog's
article.
http://www.ssicentral.com/lisrel/techdocs/HowLargeCanaStandardizedCoefficientbe.pdf
Even though the correlations are high my completely standardized
correlations are less than 1.
Questions:
What happens if one has a completely standardized coefficient of <1
but high correlations between the estimates. ?
Having a completely standardized coefficient of grater than 1 could be
a strong indicator of multicollinearity but can one also say that
having a less than 1 guarantees that there is no multicollinearity?
What is the theory behind the reasons for high estimator correlations
leading to multicollinearity. I will appreciate if there are any
references.
In my model diagnostic 2 seems fine even if the two indicators are
highly correlated in my case (the correlation between them is .97).
however the estimates are strongly correlated *,95 upwards)
Would I think that the collinearity problem is not serious?
On correcting multicollinearity:
Ed's suggested that one can do centering as has been suggested by
authors like Chronbach (1987, Psychological Bulletin) in this
listserv.
I have tried that strategy and it seemed to me that somehow in my data
it does not work I have a .96 correlation before centering and a .76
after centering. It does go down but I am not sure if this is low
enough. When I use this method diagnostic 1 still has some problems
while disgnstic 2 is fine in my model.
The other way is standardization that I found in an article by
Dunlap, William P. AND Edward R. Kemery (1988), "Effects of Predictor ...
INTERACTIONS AND Moderator Effects," Psychological Bulletin, 114 (2), 376-390
I have done log transformations to get my variable to be normal. If I
mean center this would it not affect my interpretation ?
Some clarity on this will be highly welcome. I will be very grateful.
Thanks
Sriram
--------------------------------------------------------------
To unsubscribe from SEMNET, send email to ***@bama.ua.edu
with the body of the message as: SIGNOFF SEMNET
Search the archives at http://bama.ua.edu/archives/semnet.html