Week 10 Overview

Dates 24 March 2025 - 28 MAR 2025
Reading Required: SCIU4T4 Workbook chapters 31 and 32
Recommended: Navarro and Foxcroft (2022) (Chapter 12.1 and 12.2)
Suggested: Spiegelhalter (2019) (Chapter 5)
Advanced: Morrissey and Ruxton (2018) (Download)
Lectures 10.1: Regression key concepts (15:00 min; Video)
10.2: Regression validity (14:30 min; Video)
10.3: Introduction to multiple regression (14:59 min; Video)
10.4: Regression in jamovi (min; Video)
Lecture Lecture: 26 MAR 2025 (WED) 09:00-10:00 Cottrell LT B4
Practical Using regression (Chapter 34)
Room: Cottrell 2A15
Group A: 26 MAR 2025 (WED) 10:00-13:00
Group B: 27 MAR 2025 (THU) 15:00-18:00
Help hours Brad Duthie
Room: Cottrell 2Y8
28 MAR 2025 (FRI) 14:00-16:00
Assessments Week 10 Practice quiz on Canvas

Week 10 introduces linear regression and multiple regression.

Chapter 32 uses a concrete example of fig fruit volumes across different latitudes to illustrate the key concepts of simple linear regression (i.e., regression with only one independent variable). This chapter leads with a visual interpretation of regression using a scatterplot to define the intercept, slope, and residual values in a linear regression model. This visual interpretation leads into the regression equation and how to interpret regression coefficients. Next, Chapter 32 introduces the coefficient of determination as a statistic to summarise how good any given linear model is at predicting the dependent variable. This chapter also introduces assumptions of regression coefficients, what hypothesis tests can be conducted using linear regression, and how to use linear regression to make predictions. Assumptions include that independent variable measurement is completely accurate, that the relationship between independent and dependent variables is linear, that residuals are normally distributed, and that residuals have equal variance. Sections focus on how to interpret significance of the overall regression model, along with the intercept and slope, and where to find model output in jamovi. The chapter finishes with a section on how to predict fig fruit volume from tree latitude.

Chapter 33 extends the concept of simple linear regression to multiple regression, which is defined by the existence of more than one independent variable. This chapter explains what happens when multiple independent variables are included in a linear regression model, and especially how to interpret (partial) regression coefficients in multiple regression, and why they differ from the regression coefficients calculated in simple linear regression. The chapter illustrates an output table for a multiple regression model and discusses how to interpret it correctly. The chapter then concludes by introducing a coefficient of determination that adjusts for multiple independent variables, and explains how to find and interpret the coefficient of determination in jamovi.

Chapter 34 demonstrates how to run simple and multiple linear regression models in jamovi. This chapter includes five exercises and uses a dataset that is inspired by ongoing work to understand how pyrogenic carbon is stored in soils. Pyrogenic carbon is produced when biomass is burned and becomes sequestered in soil. Exercises in this chapter test whether or not environmental variables (e.g., rainfall and soil depth) are significant for predicting the concentration of pyrogenic carbon in soils. The chapter starts with an simple linear regression model to predict pyrogenic carbon from soil depth; it demonstrates how to test linear model assumptions and report and interpret model output. This first exercise is followed by another simple linear model exercise predicting pyrogenic carbon from fire frequency before introducing an exercise on multiple regression that includes both soil depth and fire frequency as independent variables. This multiple regression exercise focuses on how to report and interpret partial regression coefficients, and contrasts the multiple regression with the first two simple lienar regressions. The chapter finishes with a large multiple regression that includes three independent variables.

References

Morrissey, Michael B., and Graeme D. Ruxton. 2018. “Multiple Regression Is Not Multiple Regressions: The Meaning of Multiple Regression and the Non-Problem of Collinearity.” Philosophy, Theory, and Practice in Biology 10 (20180709). https://doi.org/10.3998/ptpbio.16039257.0010.003.
Navarro, Danielle J, and David R Foxcroft. 2022. Learning Statistics with Jamovi. (Version 0.75). https://doi.org/10.24384/hgc3-7p15.
Spiegelhalter, David. 2019. The Art of Statistics Learning from Data. Penguin, Milton Keynes, UK.