SlideShare a Scribd company logo
Bivariate &/vs. Multivariate
• Differences between correlations, simple regression
weights & multivariate regression weights
• Patterns of bivariate & multivariate effects
• Proxy variables
• Multiple regression results to remember
It is important to discriminate among the information obtained
from ...
• a simple correlation
tells the direction and strength of the linear relationship
between two quantitative/binary variables
• a regression weight from a simple regression
tells the expected change (direction and amount) in the
criterion for a 1-unit change in the predictor
• a regression weight from a multiple regression model
tells the expected change (direction and amount) in the
criterion for a 1-unit change in that predictor, holding the
value of all the other predictors constant
Correlation
r
For a quantitative predictor
sign of r = the expected direction of change in Y as X increases
size of r = is related to the strength of that expectation
For a binary x with 0-1 coding
sign of r = tells which coded group X has higher mean Y
size of r = is related to the size of that group Y mean difference
Simple regression
y’ = bx + a raw score form
b -- raw score regression slope or coefficient
a -- regression constant or y-intercept
For a quantitative predictor
a = the expected value of y if x = 0
b = the expected direction and amount of change in the
criterion for a 1-unit increase in the
For a binary x with 0-1 coding
a = the mean of y for the group with the code value = 0
b = the mean y difference between the two coded groups
raw score regression y’ = b1x1 + b2x2 + b3x3 + a
each b
• represents the unique and independent contribution of that
predictor to the model
• for a quantitative predictor tells the expected direction and
amount of change in the criterion for a 1-unit change in that
predictor, while holding the value of all the other predictors
constant
• for a binary predictor (with unit coding -- 0,1 or 1,2, etc.), tells
direction and amount of group mean difference on the
criterion variable, while holding the value of all the other
predictors constant
a
• the expected value of the criterion if all predictors have a value
of 0
What influences the size of r, b & 
r -- bivariate correlation range = -1.00 to +1.00
-- strength of relationship with the criterion
-- sampling “problems” (e.g., range restriction)
b (raw-score regression weights range = -∞ to ∞
-- strength of relationship with the criterion
-- collinearity with the other predictors
-- differences between scale of predictor and criterion
-- sampling “problems” (e.g., range restriction)
 -- standardized regression weights range = -1.00 to +1.00
-- strength of relationship with the criterion
-- collinearity with the other predictors
-- sampling “problems” (e.g., range restriction)
Difficulties of determining “more important contributors”
-- b is not very helpful - scale differences produce b differences
--  works better, but limited by sampling variability and
measurement influences (range restriction)
Only interpret “very large”  differences as evidence that one
predictor is “more important” than another
y
x1
x2
x3
Venn diagrams representing r & b
ry,x1
ry,x2
ry,x3
y
x1
x2
x3
Remember that the b of each predictor represents the part of that
predictor shared with the criterion that is not shared with any other
predictor -- the unique contribution of that predictor to the model
bx1
bx2
bx3
Important Stuff !!! There are two different reasons that a predictor might
not be contributing to a multiple regression model...
• the variable isn’t correlated with the criterion
• the variable is correlated with the criterion, but is collinear with one or
more other predictors, and so, has no independent contribution
to the multiple regression model
y
x1
x2
x3
X1 has a substantial r with the criterion and has a substantial b
x2 has a substantial r with the criterion but has a small b because it
is collinear with x1
x3 has neither a substantial r nor substantial b
We perform both bivariate (correlation) and multivariate (multiple
regression) analyses – because they tell us different things about
the relationship between the predictors and the criterion…
Correlations (and bivariate regression weights) tell us about the
“separate” relationships of each predictor with the criterion
(ignoring the other predictors)
Multiple regression weights tell us about the relationship between
each predictor and the criterion that is unique or independent from
the other predictors in the model.
Bivariate and multivariate results for a given predictor don’t always
agree – but there is a small number of distinct patterns…
Simple correlation with the criterion
- 0 +
Multiple
regression
weight
+
0
-
Non-contributing –
probably because
colinearity with
one or more other
predictors
Non-contributing –
probably because
colinearity with
one or more other
predictors
Non-contributing –
probably because
of weak
relationship with
the criterion
Bivariate
relationship and
multivariate
contribution (to this
model) have same
sign
“Suppressor effect” –
no bivariate
relationship but
contributes (to this
model)
“Suppressor effect” –
no bivariate
relationship but
contributes (to this
model)
“Suppressor effect” –
bivariate relationship &
multivariate contribution
(to this model) have
different signs
“Suppressor effect” –
bivariate relationship &
multivariate contribution
(to this model) have
different signs
There are 5 patterns of bivariate/multivariate relationship
Bivariate
relationship and
multivariate
contribution (to this
model) have same
sign
predictor age UGPA GRE work hrs #credits
r(p) .11(.32) .45(.01) .38(.03) -.21(.06) .28(.04)
b(p) .06(.67) 1.01(.02) .002(.22) .023(.01) -.15(.03)
Bivariate & Multivariate contributions – DV = Grad GPA
Bivariate relationship and multivariate contribution (to this model)
have same sign
“Suppressor variable” – no bivariate relationship but contributes
(to this model)
“Suppressor variable” – bivariate relationship & multivariate
contribution (to this model) have different signs
Non-contributing – probably because colinearity with one or more
other predictors
Non-contributing – probably because of weak relationship with the
criterion
age
#credits
UGPA
GRE
work hrs
predictor #fish #reptiles ft2 #employees #owners
r(p) -.10(.31) .48(.01) -.28(.04) .37(.03) -.08(.54)
b(p) -.96(.03) 1.61(.42) 1.02(.02) 1.823(.01) -.65(.83)
Bivariate & Multivariate contributions – DV = Pet Quality
Non-contributing – probably because of weak relationship with the
criterion
Bivariate relationship and multivariate contribution (to this
model) have same sign
“Suppressor variable” – no bivariate relationship but
contributes (to this model)
“Suppressor variable” – bivariate relationship & multivariate
contribution (to this model) have different signs
Non-contributing – probably because of colinearity with one or
more other predictors
#fish
#reptiles
ft2
#employees
#owners
Proxy variables
Remember (again) we are not going to have experimental data!
The variables we have might be the actual causal variables influencing
this criterion, or (more likely) they might only be correlates of those
causal variables – proxy variables
Many of the “subject variables” that are very common in multivariate
modeling are of this ilk…
• is it really “sex,” “ethnicity”, “age” that are driving the criterion – or is
it all the differences in the experiences, opportunities, or other
correlates of these variables?
• is it really the “number of practices” or the things that, in turn,
produced the number of practices that were chosen?
Again, replication and convergence (trying alternative measure of
the involved constructs) can help decide if our predictors are
representing what we think the do!!
Proxy variables
In sense, proxy variables are a kind of “confounds”  because we are
attributing an effect to one variable when it might be due to another.
We can take a similar effect to understanding proxys that we do to
understanding confounds  we have to rule out specific alternative
explanations !!!
An example r gender, performance = .4 Is it really gender?
Motivation, amount of preparation & testing comfort are some
variables that have gender differences and are related to perf.
So, we run a multiple regression with all four as predictors.
If gender doesn’t contribute, then it isn’t gender but the other
variables.
If gender contributes to that model, then we know that “gender” in
the model is “the part of gender that isn’t motivation, preparation
or comfort” but we don’t know what it really is….
As we talked about last time, collinearity among the multiple predictors can
produce several patterns of bivariate-multivariate contribution. There are three
specific combinations you should be aware of (all of which are fairly rare, but
can be perplexing if they aren’t expected)…
1. Multivariate Power -- sometimes a set of predictors none of
which are significantly correlated with the criterion can be
produce a significant multivariate model (with one or more
contributing predictors)
How’s that happen?
• The error term for the multiple regression model and the test
of each predictor’s b is related to 1-R2 of the model
• Adding predictors will increase the R2 and so lower the error
term – sometimes leading to the model and 1 or more
predictors being “significant”
• This happens most often when one or more predictors have
“substantial” correlations, but the sample power is low
2. Null Washout -- sometimes a set of predictors with only one
or two significant correlations to the criterion will produce a
model that is not significant. Even worse, those significantly
correlated predictors may or may not be significant
contributors to the non-significant model
How’s that happen?
• The F-test of the model R2 really
(mathematically) tests the average
contribution of all the predictors in the model
• So, a model dominated by predictors that are not
substantially correlated with the criterion might not have a
large enough “average” contribution to be statistically
significant
• This happens most often when the sample power is low and
there are many predictors
R² / k
F = ---------------------------------
(1 - R²) / (N - k - 1)
3. Extreme collinearity -- sometimes a set of predictors all of
which are significantly correlated with the criterion can be
produce a significant multivariate model with one or more
contributing predictors
How’s that happen?
•Remember that in a multiple
regression model each predictors b
weight reflects the unique contribution
of that predictor in that model
•If the predictors are all correlated with
the criterion but are more highly
correlated with each other, each of
their “overlap” with the criterion is
shared with 1 or more other predictors
and no predictor has much unique
contribution to that very successful
(high R2) model
y
x1
x2
x3
x4

More Related Content

DOCX
Please answer all three questions . Minimal of 300 words per answe.docx
PDF
Machine Learning Algorithm - Linear Regression
PPTX
S6 w2 linear regression
PPT
Econometrics
PDF
3010l8.pdf
PPTX
SSP PRESENTATION COMPLETE ( ADVANCE ) .pptx
PPT
regression
Please answer all three questions . Minimal of 300 words per answe.docx
Machine Learning Algorithm - Linear Regression
S6 w2 linear regression
Econometrics
3010l8.pdf
SSP PRESENTATION COMPLETE ( ADVANCE ) .pptx
regression

Similar to biv_mult.ppt (20)

PPT
Introduction to Regression analysis .ppt
PPT
chapter17 introduction to regression .ppt
PPT
chapter17.ppt
PPTX
Regression
PPT
Multiple regression presentation
PPT
A presentation for Multiple linear regression.ppt
PDF
Group 5 - Regression Analysis.pdf
PDF
Quantitative Methods for Lawyers - Class #19 - Regression Analysis - Part 2
PDF
Research method ch09 statistical methods 3 estimation np
PPT
Chapter 4 - multiple regression
PPT
Lecture 4
DOCX
Exercise 29Calculating Simple Linear RegressionSimple linear reg.docx
PPT
Multivariate Linear Regression.ppt
PPT
TTests.ppt
PPTX
Regression analysis
PPTX
Regression analysis
PPTX
Regression
PPT
week9b.pptfgfgggggggggggggggggggggggggggggggggggg
PPT
Regression and Co-Relation
DOCX
62083750 multiple-regression
Introduction to Regression analysis .ppt
chapter17 introduction to regression .ppt
chapter17.ppt
Regression
Multiple regression presentation
A presentation for Multiple linear regression.ppt
Group 5 - Regression Analysis.pdf
Quantitative Methods for Lawyers - Class #19 - Regression Analysis - Part 2
Research method ch09 statistical methods 3 estimation np
Chapter 4 - multiple regression
Lecture 4
Exercise 29Calculating Simple Linear RegressionSimple linear reg.docx
Multivariate Linear Regression.ppt
TTests.ppt
Regression analysis
Regression analysis
Regression
week9b.pptfgfgggggggggggggggggggggggggggggggggggg
Regression and Co-Relation
62083750 multiple-regression
Ad

More from PerumalPitchandi (20)

PPT
Workplaces Ethics_CCGA_Human_Factors.ppt
PPT
Analysis technologies - day3 slides Lecture notesppt
PPT
Introduction to computer networks lecture
PPTX
20IT204-Computer Organization and Architecture-Lecture 1.pptx
PPTX
Introduction to Software Quality Metrics
PPTX
Introduction to Test Automation Notes.pptx
PPT
Agile Methodology-extreme programming-23.07.2020.ppt
PPTX
Lecture Notes on Recommender System Introduction
PPTX
22ADE002 – Business Analytics- Module 1.pptx
PDF
ppt_ids-data science.pdf
PPT
ANOVA Presentation.ppt
PPTX
Data Science Intro.pptx
PPT
Descriptive_Statistics_PPT.ppt
PPT
SW_Cost_Estimation.ppt
PPT
CostEstimation-1.ppt
PPT
20IT204-COA-Lecture 18.ppt
PPTX
20IT204-COA- Lecture 17.pptx
PPTX
Capability Maturity Model (CMM).pptx
PPTX
Comparison_between_Waterfall_and_Agile_m (1).pptx
PPTX
Introduction to Data Science.pptx
Workplaces Ethics_CCGA_Human_Factors.ppt
Analysis technologies - day3 slides Lecture notesppt
Introduction to computer networks lecture
20IT204-Computer Organization and Architecture-Lecture 1.pptx
Introduction to Software Quality Metrics
Introduction to Test Automation Notes.pptx
Agile Methodology-extreme programming-23.07.2020.ppt
Lecture Notes on Recommender System Introduction
22ADE002 – Business Analytics- Module 1.pptx
ppt_ids-data science.pdf
ANOVA Presentation.ppt
Data Science Intro.pptx
Descriptive_Statistics_PPT.ppt
SW_Cost_Estimation.ppt
CostEstimation-1.ppt
20IT204-COA-Lecture 18.ppt
20IT204-COA- Lecture 17.pptx
Capability Maturity Model (CMM).pptx
Comparison_between_Waterfall_and_Agile_m (1).pptx
Introduction to Data Science.pptx
Ad

Recently uploaded (20)

PDF
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
PPTX
UNIT-1 - COAL BASED THERMAL POWER PLANTS
PPTX
CH1 Production IntroductoryConcepts.pptx
PPTX
KTU 2019 -S7-MCN 401 MODULE 2-VINAY.pptx
PPTX
Internet of Things (IOT) - A guide to understanding
DOCX
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
DOCX
573137875-Attendance-Management-System-original
PPT
Project quality management in manufacturing
PPTX
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
PDF
Arduino robotics embedded978-1-4302-3184-4.pdf
PPTX
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
PDF
Embodied AI: Ushering in the Next Era of Intelligent Systems
PPTX
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
PPTX
Foundation to blockchain - A guide to Blockchain Tech
PDF
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
PPTX
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
PPTX
Welding lecture in detail for understanding
PPTX
Geodesy 1.pptx...............................................
PPTX
additive manufacturing of ss316l using mig welding
PPTX
Lesson 3_Tessellation.pptx finite Mathematics
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
UNIT-1 - COAL BASED THERMAL POWER PLANTS
CH1 Production IntroductoryConcepts.pptx
KTU 2019 -S7-MCN 401 MODULE 2-VINAY.pptx
Internet of Things (IOT) - A guide to understanding
ASol_English-Language-Literature-Set-1-27-02-2023-converted.docx
573137875-Attendance-Management-System-original
Project quality management in manufacturing
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
Arduino robotics embedded978-1-4302-3184-4.pdf
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
Embodied AI: Ushering in the Next Era of Intelligent Systems
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
Foundation to blockchain - A guide to Blockchain Tech
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
Welding lecture in detail for understanding
Geodesy 1.pptx...............................................
additive manufacturing of ss316l using mig welding
Lesson 3_Tessellation.pptx finite Mathematics

biv_mult.ppt

  • 1. Bivariate &/vs. Multivariate • Differences between correlations, simple regression weights & multivariate regression weights • Patterns of bivariate & multivariate effects • Proxy variables • Multiple regression results to remember
  • 2. It is important to discriminate among the information obtained from ... • a simple correlation tells the direction and strength of the linear relationship between two quantitative/binary variables • a regression weight from a simple regression tells the expected change (direction and amount) in the criterion for a 1-unit change in the predictor • a regression weight from a multiple regression model tells the expected change (direction and amount) in the criterion for a 1-unit change in that predictor, holding the value of all the other predictors constant
  • 3. Correlation r For a quantitative predictor sign of r = the expected direction of change in Y as X increases size of r = is related to the strength of that expectation For a binary x with 0-1 coding sign of r = tells which coded group X has higher mean Y size of r = is related to the size of that group Y mean difference
  • 4. Simple regression y’ = bx + a raw score form b -- raw score regression slope or coefficient a -- regression constant or y-intercept For a quantitative predictor a = the expected value of y if x = 0 b = the expected direction and amount of change in the criterion for a 1-unit increase in the For a binary x with 0-1 coding a = the mean of y for the group with the code value = 0 b = the mean y difference between the two coded groups
  • 5. raw score regression y’ = b1x1 + b2x2 + b3x3 + a each b • represents the unique and independent contribution of that predictor to the model • for a quantitative predictor tells the expected direction and amount of change in the criterion for a 1-unit change in that predictor, while holding the value of all the other predictors constant • for a binary predictor (with unit coding -- 0,1 or 1,2, etc.), tells direction and amount of group mean difference on the criterion variable, while holding the value of all the other predictors constant a • the expected value of the criterion if all predictors have a value of 0
  • 6. What influences the size of r, b &  r -- bivariate correlation range = -1.00 to +1.00 -- strength of relationship with the criterion -- sampling “problems” (e.g., range restriction) b (raw-score regression weights range = -∞ to ∞ -- strength of relationship with the criterion -- collinearity with the other predictors -- differences between scale of predictor and criterion -- sampling “problems” (e.g., range restriction)  -- standardized regression weights range = -1.00 to +1.00 -- strength of relationship with the criterion -- collinearity with the other predictors -- sampling “problems” (e.g., range restriction) Difficulties of determining “more important contributors” -- b is not very helpful - scale differences produce b differences --  works better, but limited by sampling variability and measurement influences (range restriction) Only interpret “very large”  differences as evidence that one predictor is “more important” than another
  • 7. y x1 x2 x3 Venn diagrams representing r & b ry,x1 ry,x2 ry,x3
  • 8. y x1 x2 x3 Remember that the b of each predictor represents the part of that predictor shared with the criterion that is not shared with any other predictor -- the unique contribution of that predictor to the model bx1 bx2 bx3
  • 9. Important Stuff !!! There are two different reasons that a predictor might not be contributing to a multiple regression model... • the variable isn’t correlated with the criterion • the variable is correlated with the criterion, but is collinear with one or more other predictors, and so, has no independent contribution to the multiple regression model y x1 x2 x3 X1 has a substantial r with the criterion and has a substantial b x2 has a substantial r with the criterion but has a small b because it is collinear with x1 x3 has neither a substantial r nor substantial b
  • 10. We perform both bivariate (correlation) and multivariate (multiple regression) analyses – because they tell us different things about the relationship between the predictors and the criterion… Correlations (and bivariate regression weights) tell us about the “separate” relationships of each predictor with the criterion (ignoring the other predictors) Multiple regression weights tell us about the relationship between each predictor and the criterion that is unique or independent from the other predictors in the model. Bivariate and multivariate results for a given predictor don’t always agree – but there is a small number of distinct patterns…
  • 11. Simple correlation with the criterion - 0 + Multiple regression weight + 0 - Non-contributing – probably because colinearity with one or more other predictors Non-contributing – probably because colinearity with one or more other predictors Non-contributing – probably because of weak relationship with the criterion Bivariate relationship and multivariate contribution (to this model) have same sign “Suppressor effect” – no bivariate relationship but contributes (to this model) “Suppressor effect” – no bivariate relationship but contributes (to this model) “Suppressor effect” – bivariate relationship & multivariate contribution (to this model) have different signs “Suppressor effect” – bivariate relationship & multivariate contribution (to this model) have different signs There are 5 patterns of bivariate/multivariate relationship Bivariate relationship and multivariate contribution (to this model) have same sign
  • 12. predictor age UGPA GRE work hrs #credits r(p) .11(.32) .45(.01) .38(.03) -.21(.06) .28(.04) b(p) .06(.67) 1.01(.02) .002(.22) .023(.01) -.15(.03) Bivariate & Multivariate contributions – DV = Grad GPA Bivariate relationship and multivariate contribution (to this model) have same sign “Suppressor variable” – no bivariate relationship but contributes (to this model) “Suppressor variable” – bivariate relationship & multivariate contribution (to this model) have different signs Non-contributing – probably because colinearity with one or more other predictors Non-contributing – probably because of weak relationship with the criterion age #credits UGPA GRE work hrs
  • 13. predictor #fish #reptiles ft2 #employees #owners r(p) -.10(.31) .48(.01) -.28(.04) .37(.03) -.08(.54) b(p) -.96(.03) 1.61(.42) 1.02(.02) 1.823(.01) -.65(.83) Bivariate & Multivariate contributions – DV = Pet Quality Non-contributing – probably because of weak relationship with the criterion Bivariate relationship and multivariate contribution (to this model) have same sign “Suppressor variable” – no bivariate relationship but contributes (to this model) “Suppressor variable” – bivariate relationship & multivariate contribution (to this model) have different signs Non-contributing – probably because of colinearity with one or more other predictors #fish #reptiles ft2 #employees #owners
  • 14. Proxy variables Remember (again) we are not going to have experimental data! The variables we have might be the actual causal variables influencing this criterion, or (more likely) they might only be correlates of those causal variables – proxy variables Many of the “subject variables” that are very common in multivariate modeling are of this ilk… • is it really “sex,” “ethnicity”, “age” that are driving the criterion – or is it all the differences in the experiences, opportunities, or other correlates of these variables? • is it really the “number of practices” or the things that, in turn, produced the number of practices that were chosen? Again, replication and convergence (trying alternative measure of the involved constructs) can help decide if our predictors are representing what we think the do!!
  • 15. Proxy variables In sense, proxy variables are a kind of “confounds”  because we are attributing an effect to one variable when it might be due to another. We can take a similar effect to understanding proxys that we do to understanding confounds  we have to rule out specific alternative explanations !!! An example r gender, performance = .4 Is it really gender? Motivation, amount of preparation & testing comfort are some variables that have gender differences and are related to perf. So, we run a multiple regression with all four as predictors. If gender doesn’t contribute, then it isn’t gender but the other variables. If gender contributes to that model, then we know that “gender” in the model is “the part of gender that isn’t motivation, preparation or comfort” but we don’t know what it really is….
  • 16. As we talked about last time, collinearity among the multiple predictors can produce several patterns of bivariate-multivariate contribution. There are three specific combinations you should be aware of (all of which are fairly rare, but can be perplexing if they aren’t expected)… 1. Multivariate Power -- sometimes a set of predictors none of which are significantly correlated with the criterion can be produce a significant multivariate model (with one or more contributing predictors) How’s that happen? • The error term for the multiple regression model and the test of each predictor’s b is related to 1-R2 of the model • Adding predictors will increase the R2 and so lower the error term – sometimes leading to the model and 1 or more predictors being “significant” • This happens most often when one or more predictors have “substantial” correlations, but the sample power is low
  • 17. 2. Null Washout -- sometimes a set of predictors with only one or two significant correlations to the criterion will produce a model that is not significant. Even worse, those significantly correlated predictors may or may not be significant contributors to the non-significant model How’s that happen? • The F-test of the model R2 really (mathematically) tests the average contribution of all the predictors in the model • So, a model dominated by predictors that are not substantially correlated with the criterion might not have a large enough “average” contribution to be statistically significant • This happens most often when the sample power is low and there are many predictors R² / k F = --------------------------------- (1 - R²) / (N - k - 1)
  • 18. 3. Extreme collinearity -- sometimes a set of predictors all of which are significantly correlated with the criterion can be produce a significant multivariate model with one or more contributing predictors How’s that happen? •Remember that in a multiple regression model each predictors b weight reflects the unique contribution of that predictor in that model •If the predictors are all correlated with the criterion but are more highly correlated with each other, each of their “overlap” with the criterion is shared with 1 or more other predictors and no predictor has much unique contribution to that very successful (high R2) model y x1 x2 x3 x4