This lesson begins the discussion of building models - regression models (linear regression, logistic regression, generalized linear models) and ANOVA (one-way, two-way, mixed) and others…
We will be referring to the PPT/PDF slides: (click on the links to download the files)
ResearchRoundtableBuildRegressionModel_final14Nov08.ppt
https://github.com/melindahiggins2000/N736Fall2017/raw/master/ResearchRoundtableBuildRegressionModel_final14Nov08.ppt
ResearchRoundtableBuildRegressionModel_final14Nov08.pdf
https://github.com/melindahiggins2000/N736Fall2017/raw/master/ResearchRoundtableBuildRegressionModel_final14Nov08.pdf
Assume that the terms are “linearly”" related to the outcome - that relationship is linear (either directly or through transformations)
The “slope” or “relationship” between X & Y is assumed to be “constant” / “consistent” across all levels of X
We assume that Y was observed for all levels of X and that it was observed “evenly” (“balanced” experimental design) - “Anscome Quartet”
The independent variables are independent of one another (no multicollinearity) – although some “mild/minor” correlation may be tolerated. Check the Variance Inflation Factor (VIF) and/or the Tolerance (TOL), which is the inverse of the VIF.
Any case which has “missing” data on any of the IVs or DV will be eliminated from the analysis (listwise deletion).
The intercept and all coefficients for the IVs are “Fixed.” [“Random Coefficient Models” to be discussed later.]
We assume that the variance of Y is constant across all levels of X (assumption of homogeneity of variance) - important for regression and ANOVA (e.g. “Levene’s test”)
The “error” term is additive (not multiplicative)
Assume “errors” (“residuals”) are normally distributed, identical and independent ~iid N(mu,sigma).
[next lesson - discussion of Type I, II and III sums of squares - differences in SPSS, SAS and R]
Copyright © Melinda Higgins, Ph.D.. All contents under (CC) BY-NC-SA license, unless otherwise noted.