-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathweek12-part2.Rmd
More file actions
218 lines (150 loc) · 7.77 KB
/
week12-part2.Rmd
File metadata and controls
218 lines (150 loc) · 7.77 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
---
title: "Defining and Detecting Multicollinearity"
author: "STAT 021 with Prof Suzy"
institute: "Swarthmore College"
output:
xaringan::moon_reader:
# css: ["default", "assets/sydney-fonts.css", "assets/sydney.css"]
self_contained: false # if true, fonts will be stored locally
# includes:
# in_header: "assets/mathjax-equation-numbers.html"
nature:
# beforeInit: ["assets/remark-zoom.js", "https://platform.twitter.com/widgets.js"]
highlightStyle: github
highlightLines: true
countIncrementalSlides: false
ratio: '16:9' # alternatives '16:9' or '4:3' or others e.g. 13:9
navigation:
scroll: false # disable slide transitions by scrolling
includes:
in_header: mystyles.sty
---
```{r setup_pres, include=FALSE, echo=FALSE}
#devtools::install_github("ropenscilabs/icon")
#devtools::session_info('rmarkdown')
knitr::opts_chunk$set(message = FALSE) # include this if you don't want markdown to knit messages
knitr::opts_chunk$set(warning = FALSE) # include this if you don't want markdown to knit warnings
rm(list=ls())
library('tidyverse')
library('gridExtra')
library('broom')
library('cowplot')
library("RefManageR")
library("DT")
library('scatterplot3d')
```
```{css, echo=FALSE}
pre {
background: #FFBB33;
max-width: 100%;
overflow-x: scroll;
}
.scroll-output {
height: 70%;
overflow-y: scroll;
}
.scroll-small {
height: 50%;
overflow-y: scroll;
}
.red{color: #ce151e;}
.green{color: #26b421;}
.blue{color: #426EF0;}
```
## Multicollinearity
There are a lot of important c-words in statistics. Especially confusing can be the following: collinearity, correlation, and covariance. But, somewhat intuitively, there is a relationship among these three terms.
Recall the definitions of correlation and covariance for any two random variables, say, $X$ and $Y$:
$$Cov(X,Y) = E[(X -E[X])(Y-E[Y])] = \dots = E[XY] - E[X]E[Y]$$
$$Cor(X,Y) = \frac{Cov(X,Y)}{\sqrt{Var(X)Var(Y)}}$$
--
Also, recall the fact that .blue[if $X$ and $Y$ are independent, then $Cov(X,Y)=0$ and therefore $X$ and $Y$ are uncorrelated.]
**But** if $X$ and $Y$ are uncorrelated, then it is still possible for $X$ and $Y$ to be dependent or independent.
---
## Multicollinearity
So now we see that correlation is just a standardized version of the covariance between two variables. (Standardized in the sense that it will always be between the interval $[-1,1]$.)
**Q:** What is multicollinearity?
--
This is a term specific to MLR models that describes the statistical phenomenon in which two or more predictor variables are highly correlated with each other.
In other words, this means that two predictor variables, say, $x_1$ and $x_2$, are collinear if $Cor(x_1, x_2) \approx \pm 1$.
- This means that we could predict the values of one from the other in a SLR model!
- This is **not** a problem for .blue[estimation] or .blue[prediction] with a MLR.
- This **is** a problem however that inflates the estimated variances of our regression coefficients and therefore the individual test of significant slope parameters. Thus it can be a problem for .red[inference].<sup>[5]</sup>
- Outliers can have a big impact on the collinearity of a pair of variables.<sup>[5]</sup>
---
## Multicollinearity Example
### Public School SAT data
.scroll-output[Again let's investigate the MLR model with four predictor variables that we built to predict SAT scores of public schools. Does it seem like any of the predictor variables might be correlated with each other?
```{r}
SAT_data <- read_table2("Data/sat_data.txt", col_names=FALSE, cols(col_character(), col_double(), col_double(), col_double(), col_double(), col_double(), col_double(), col_double()))
colnames(SAT_data) = c("State", "PerPupilSpending", "StuTeachRatio", "Salary", "PropnStu", "SAT_verbal", "SAT_math", "SAT_tot")
SAT_data <- SAT_data %>% mutate(prop_taking_SAT = PropnStu/100) %>% select(-PropnStu)
SAT_data %>% select(c(PerPupilSpending, StuTeachRatio, Salary, prop_taking_SAT)) %>% plot
```
]
---
## Multicollinearity Example
### Public School SAT data
.scroll-output[Again let's investigate the MLR model with four predictor variables that we built to predict SAT scores of public schools. Does it seem like any of the predictor variables might be correlated with each other?
```{r}
MLR_SAT4 <- lm(SAT_tot ~ PerPupilSpending +
StuTeachRatio +
Salary +
prop_taking_SAT, data = SAT_data)
summary(MLR_SAT4)
```
]
---
## Multicollinearity Example
### Public School SAT data
We can estimate the correlations among each pair of variables (without risking multiple testing issues):
.scroll-small[
```{r}
cor(SAT_data$PerPupilSpending, SAT_data$StuTeachRatio)
cor(SAT_data$PerPupilSpending, SAT_data$Salary)
cor(SAT_data$PerPupilSpending, SAT_data$prop_taking_SAT)
cor(SAT_data$StuTeachRatio, SAT_data$Salary)
cor(SAT_data$StuTeachRatio, SAT_data$prop_taking_SAT)
cor(SAT_data$Salary, SAT_data$prop_taking_SAT)
```
Note: The estimated correlation between two variables is not affected by whether or not those variables are standardized because correlation itself is already standardized.
]
---
## Multicollinearity
When we see evidence of collinear predictor variables, it's a good idea to re-visit which variables you want to include in the model .red[because] collinearity affects the variance of our predictors and therefore affects the conclusion of the individual t-tests!<sup>[4]</sup>
**Q:** Do we need to be concerned about multiple testing with interaction terms? with collinearity?
--
**A:** We do need to keep an eye on multiple tests issues when we're determining which main effects and interaction effects to include in the model (these are questions of statistical significance). The problem of multiple testing is not a concern however, in estimation/prediction problems therefore it is not a major concern when checking for collinearity (all we're doing is getting estimates for different correlations).
**Note:** Interaction terms are structurally collinear, which we can't do much about.
---
## Multicollinearity
### Remedies
A couple of simple remedies for collinearity are
- Try collecting more data to see if the collinearity is dues to insufficient data;
- Reduce the number of variables in your model so that none of them are collinear;
- Other regression methods such as *ridge regression* (you don't need to know what this is for my class).<sup>[5]</sup>
---
## Multicollinearity
### Public School SAT data
Finally we can get a better model for the SAT data by adjusting which variables we include:
.scroll-small[
```{r echo=FALSE}
MLR_SAT_good <- lm(SAT_tot ~ PerPupilSpending +
StuTeachRatio +
prop_taking_SAT, data = SAT_data)
summary(MLR_SAT_good)
```
]
---
## Interaction terms and multicollinearity
Broadly speaking, there are two main types of collinearity:
1. Structural multicollinearity which occurs when we create a model term using other terms in the model (e.g. including interaction terms). This is a (unfortunately unavoidable) byproduct of the model that we specify.
2. Data multicollinearity which occurs when the data itself for different predictor variables are highly correlated. Some ways we can address this data-inherent multicollinearity is to collect more data or see if it makes sense to drop a variable from the model.
.footnote[Source: https://statisticsbyjim.com/regression/multicollinearity-in-regression-analysis/]
---
## Multicollinearity
### Some things to think about
- What are the effects of incorporating interaction terms in your linear model?
- What are the effects of severe multicollinearity among some predictor variables in your model?
***
### Reading along in your textbook
Chapter 3 Sections 10 and 11 and Chapter 9