Biology Forums - Study Force

Science-Related Homework Help Computer Studies Topic started by: seeb1999 on Aug 24, 2019



Title: List the five assumptions made in linear regressions and select one to discuss in depth.
Post by: seeb1999 on Aug 24, 2019
List the five assumptions made in linear regressions and select one to discuss in depth.


Title: List the five assumptions made in linear regressions and select one to discuss in depth.
Post by: binva on Aug 24, 2019
1. Linearity. This assumption states that the relationship between the response variable and the explanatory variables is linear. That is, the expected value of the response variable is a straight-line function of each explanatory variable while holding all other explanatory variables fixed. Also, the slope of the line does not depend on the values of the other variables. It also implies that the effects of different explanatory variables on the expected value of the response variable are additive in nature.
2. Independence (of errors). This assumption states that the errors of the response variable are uncorrelated with each other. This independence of the errors is weaker
than actual statistical independence, which is a stronger condition and is often not needed for linear regression analysis.
3. Normality (of errors). This assumption states that the errors of the response variable are normally distributed. That is, they are supposed to be totally random and should not represent any nonrandom patterns.
4. Constant variance (of errors). This assumption, also called homoscedasticity, states that the response variables have the same variance in their error regardless of the values of the explanatory variables. In practice, this assumption is invalid if the response variable varies over a wide enough range/scale.
5. Multicollinearity. This assumption states that the explanatory variables are not correlated (i.e., do not replicate the same but provide a different perspective of the information needed for the model). Multicollinearity can be triggered by having two or more perfectly correlated explanatory variables presented to the model (e.g., if the same explanatory variable is mistakenly included in the model twice, one with a slight transformation of the same variable). A correlation-based data assessment usually catches this error.