--- title: "lecture12" output: html_document --- ```{r} # Assessing the fit of a mean # Example to Slide 4 # The data sample politeness <- c(10, 10, 20, 20, 30, 30, 40, 40, 50, 50, 60, 60) likability <- c(3, 8, 9, 10, 7, 9, 9, 11, 11, 12, 10, 13) # The model m <- mean(likability) m # Plotting plot(politeness, likability) abline(h = m) # Residual Standard Error sse <- sum((likability - mean(likability))**2) df <- (length(politeness) - 1) s <- sqrt(sse / df) s ``` ```{r} # Assessing the fit of a regression line # Example to Slide 5 # The data sample politeness <- c(10, 10, 20, 20, 30, 30, 40, 40, 50, 50, 60, 60) likability <- c(3, 8, 9, 10, 7, 9, 9, 11, 11, 12, 10, 13) # Fitting the linear model m <- lm(likability ~ politeness) m # Plotting plot(politeness, likability) abline(m) # Residual Standard Error sse <- sum((likability - predict(m, data.frame(politeness)))**2) df <- (length(politeness) - 1) s <- sqrt(sse / df) s # Predict Likability from Politeness = 45 predict(m, data.frame(politeness = 45)) ``` ```{r} # Slide 11 height <- c(150, 150, 160, 160, 160, 175, 175, 180, 180) weight <- c(50, 60, 65, 90, 80, 85, 75, 80, 90) plot(height, weight) m <- lm(weight ~ height) abline(m) m$coefficients ``` ```{r} # Slide 11 # Assessing the significance of the predictor # Example to Slide 15 # The data sample politeness <- c(10, 10, 20, 20, 30, 30, 40, 40, 50, 50, 60, 60) likability <- c(3, 8, 9, 10, 7, 9, 9, 11, 11, 12, 10, 13) # Building the linear model m <- lm(likability ~ politeness) # Print the statistics. # It will tell you that the coefficient for politeness (our b1) is 0.1086, # and that it is significantly different from 0. The t value is 3.541, # and the p value 0.005593 (two-tailed test). summary(m) ``` ```{r} # Slide 18 groups <- c(0, 0, 0, 0, 0, 1, 1, 1, 1, 1) data <- c(5, 3, 2, 4, 2, 5, 8, 9, 10, 8) # Running a Two-Sample t-Test t.test(data[6:10], data[1:5], var.equal = T) # Building a Linear model m <- lm(data ~ groups) # Plotting plot(groups, data) abline(m) # Printing the statistics # You can see that the mean of group0 is 3.2 (the intercept) and that the coefficient for the group variable # is 4.8. Thus, the mean of group1 is 3.2 + 4.8 = 8.0. Moreover, the t-Test yields the same t- and p value as the # one using the t.test() procedure. summary(m) ``` ```{r} # Slide 20 groups <- c(0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 2, 2, 2, 2, 2) data <- c(5, 3, 2, 4, 2, 5, 8, 9, 10, 8, 6, 8, 9, 8, 12) # Running the ANOVA a <- aov(data ~ groups) summary(a) # Building a Linear model m <- lm(data ~ groups) # Plotting plot(groups, data) abline(m) # Printing the statistics # You can see that the F value and the p value of the linear regression model # is equal to the F and p value from the anova. summary(m) ```