Professional Documents
Culture Documents
Chapter 16
Partial Correlation and Multiple
Regression and Correlation
In This Presentation
Partial correlations
Multiple regression
Using the multiple regression line to predict Y
Multiple correlation coefficient (R2)
Limitations of multiple regression and
correlation
Introduction
Multiple Regression and Correlation allow us
to:
1.
2.
3.
Partial Correlation
Partial Correlation measures the correlation
Partial Correlation
Note the subscripts in the symbol for a partial
correlation coefficient:
rxyz
which indicates that the correlation coefficient is for X
and Y controlling for Z
Partial Correlation
Example
The table below lists husbands hours of housework per week (Y),
number of children (X), and husbands years of education (Z) for a
sample of 12 dual-career households
Partial Correlation
Example
A correlation matrix appears below
The bivariate (zero-order) correlation between husbands
housework and number of children is +0.50
This indicates a positive relationship
Partial Correlation
Example
Calculating the partial (first-order) correlation between
husbands housework and number of children controlling for
husbands years of education yields +0.43
Partial Correlation
Example
Comparing the bivariate correlation (+0.50) to
the partial correlation (+0.43) finds little
change
The relationship between number of children
and husbands housework controlling for
husbands education has not changed
Therefore, we have evidence of a direct
relationship
Multiple Regression
Previously, the bivariate regression equation was:
Multiple Regression
Y = a + b1X1 + b2X2
Notation
a is the Y intercept, where the regression line crosses the
Y axis
b1 is the partial slope for X1 on Y
b1 indicates the change in Y for one unit change in X1,
controlling for X2
b2 is the partial slope for X2 on Y
b2 indicates the change in Y for one unit change in X2,
controlling for X1
r YX1
r YX2
Sig. (1-tailed)
Daily calorie
intake
People who
read (%)
1.000
.776
.869
.776
.869
1.000
.682
.682
1.000
.000
.000
.000
.000
.
.000
.000
.
74
74
74
74
74
74
74
74
74
r X1X2
Model
1
(Constant)
People who read (%)
Daily calorie intake
Unstandardized
Coefficients
B
Std. Error
25.838
2.882
.315
.034
.007
.001
Standardized
Coefficients
Beta
.636
.342
t
8.964
9.202
4.949
Sig.
.000
.000
.000
Zero-order
Correlations
Partial
.869
.776
.738
.506
Part
.465
.250
Collinearity Statistics
Tolerance
VIF
.535
.535
1.868
1.868
R
.905a
R Square
.818
Adjusted
R Square
.813
Std. Error of
the Estimate
4.948
R Square
Change
.818
F Change
159.922
df1
df2
2
71
Sig. F Change
.000
Regression
Residual
Total
Sum of
Squares
7829.451
1738.008
9567.459
df
2
71
73
Mean Square
3914.726
24.479
F
159.922
Sig.
.000a
Model
1
(Constant)
Daily calorie intake
People who read (%)
Unstandardized
Coefficients
B
Std. Error
25.838
2.882
.007
.001
.315
.034
Standardized
Coefficients
Beta
.342
.636
t
8.964
4.949
9.202
Sig.
.000
.000
.000
Zero-order
Correlations
Partial
.776
.869
.506
.738
Part
.250
.465
Limitations
Multiple regression and correlation are among the most powerful
techniques available to researchers. But powerful techniques have
high demands.
These techniques require:
Every variable is measured at the interval-ratio level
Each independent variable has a linear relationship with the
dependent variable
Independent variables do not interact with each other
Independent variables are uncorrelated with each other
When these requirements are violated (as they often are), these
techniques will produce biased and/or inefficient estimates. There
are more advanced techniques available to researchers that can
correct for violations of these requirements. Such techniques are
beyond the scope of this text.