Entry
Reader's guide
Entries A-Z
Subject index
Multicollinearity
Collinearity between two INDEPENDENT VARIABLES or multicollinearity between multiple independent variables in LINEAR REGRESSION analysis means that there are linear relations between these variables. In a vector model, in which variables are represented as vectors, exact collinearity would mean that two vectors lie on the same line or, more generally, for k variables, that the vectors lie in a subspace of a dimension less than k (i.e., that one of the vectors is a linear combination of the others). The result of such an exact linear relation between variables X would be that the cross-product matrix X′X is singular (i.e., its determinant would be equal to zero), so that it would have no inverse. In practice, an exact linear relationship rarely occurs, but the interdependence of social phenomena may result in “approximate” linear relationships. This phenomenon is known as multicollinearity.
In the analysis of interdependence, such as PRINCIPAL COMPONENTS and FACTOR ANALYSIS, multicollinearity is blissful because dimension reduction is the objective, whereas in the analysis of dependence structures, such as MULTIPLE REGRESSION analysis, multicollinearity among the independent variables is a problem. We will treat it here in the context of regression analysis.
Consequences
Imagine an investigation into the causes of wellbeing of unemployed workers (measured as a scale from 0 to 100). A great number of explanatory variables will be taken into account, such as annual income (the lower the annual income, the lower the well-being), duration of unemployment (the higher the number of weeks of unemployment, the lower the well-being), and others. Now, some of these independent variables will be so highly correlated that conclusions from a multiple regression analysis will be affected. High multicollinearity is undesirable in a multicausal model because standard errors become much larger and CONFIDENCE INTERVALS much wider, thus affecting SIGNIFICANCE TESTING and STATISTICAL INFERENCE, even though the ORDINARY LEAST SQUARES (OLS) estimator still is unbiased. Like variances and covariances, multicollinearity is a sample property. This means analyses may be unreliable and sensitive toafew outliers or extreme cases.
The problem of correlated independent variables in a causal model is an old sore in social scientific research. In multiple regression analysis, it is referred to as multicollinearity (the predictors in a vector model almost overlap in a straight line), whereas in ANALYSIS OF VARIANCE and COVARIANCE, one tends to use the term nonorthogonality (the factors of the factor-response model are not perpendicular to each other). Multicollinearity and nonorthogonality are, in fact, synonyms.
Let us emphasize that strange things may happen in an analysis with high multicollinearity. Kendall and Stuart (1948) provided an extreme example of one dependent variable Y and two independent variables X1 and X2, in which the correlations Y −X1 and Y −X2 are weak (one of them even zero) and the correlation X1 − X2 is very strong (0.90, high multicollinearity), whereas the multiple correlation coefficient Ry·12 is approximately 1. This classic example shows a classic consequence of multicollinearity: insignificant results of testing individual parameters but high R-SQUARED values. The standardized coefficients (a direct function of sample variances), which normally cannot be greater than 1 because their squared values express a proportion of explained variance, exceed 2; consequently, both variables appear to explain each more than 400%, with only 100% to be explained!
...
- Analysis of Variance
- Association and Correlation
- Association
- Association Model
- Asymmetric Measures
- Biserial Correlation
- Canonical Correlation Analysis
- Correlation
- Correspondence Analysis
- Intraclass Correlation
- Multiple Correlation
- Part Correlation
- Partial Correlation
- Pearson's Correlation Coefficient
- Semipartial Correlation
- Simple Correlation (Regression)
- Spearman Correlation Coefficient
- Strength of Association
- Symmetric Measures
- Basic Qualitative Research
- Basic Statistics
- F Ratio
- N(n)
- t-Test
- X¯
- Y Variable
- z-Test
- Alternative Hypothesis
- Average
- Bar Graph
- Bell-Shaped Curve
- Bimodal
- Case
- Causal Modeling
- Cell
- Covariance
- Cumulative Frequency Polygon
- Data
- Dependent Variable
- Dispersion
- Exploratory Data Analysis
- Frequency Distribution
- Histogram
- Hypothesis
- Independent Variable
- Measures of Central Tendency
- Median
- Null Hypothesis
- Pie Chart
- Regression
- Standard Deviation
- Statistic
- Causal Modeling
- Discourse/Conversation Analysis
- Econometrics
- Epistemology
- Ethnography
- Evaluation
- Event History Analysis
- Experimental Design
- Factor Analysis and Related Techniques
- Feminist Methodology
- Generalized Linear Models
- Historical/Comparative
- Interviewing in Qualitative Research
- Latent Variable Model
- Life History/Biography
- Log-Linear Models (Categorical Dependent Variables)
- Longitudinal Analysis
- Mathematics and Formal Models
- Measurement Level
- Measurement Testing and Classification
- Multilevel Analysis
- Multiple Regression
- Qualitative Data Analysis
- Sampling in Qualitative Research
- Sampling in Surveys
- Scaling
- Significance Testing
- Simple Regression
- Survey Design
- Time Series
- ARIMA
- Box-Jenkins Modeling
- Cointegration
- Detrending
- Durbin-Watson Statistic
- Error Correction Models
- Forecasting
- Granger Causality
- Interrupted Time-Series Design
- Intervention Analysis
- Lag Structure
- Moving Average
- Periodicity
- Serial Correlation
- Spectral Analysis
- Time-Series Cross-Section (TSCS) Models
- Time-Series Data (Analysis/Design)
- Trend Analysis
- Loading...
Get a 30 day FREE TRIAL
-
Watch videos from a variety of sources bringing classroom topics to life
-
Read modern, diverse business cases
-
Explore hundreds of books and reference titles
Sage Recommends
We found other relevant content for you on other Sage platforms.
Have you created a personal profile? Login or create a profile so that you can save clips, playlists and searches