A simple linear regression is a linear regression in which there is only one covariate (predictor variable).
Simple linear regression is used in situations to evaluate the linear relationship between two variables. One example could be the relationship between muscle strength and lean body mass. Another way to put it is that simple linear regression is used to develop an equation by which we can predict or estimate a dependent variable given an independent variable.
The regression equation is given by
Where is the dependent variable, is the y intercept, is the gradient or slope of the line and is independent variable.
The linear relationship between the two variables (i.e. dependent and independent) can be measured using a correlation coefficient e.g. the Pearson product moment correlation coefficient.
Estimating the Regression Line
The SLR(simple linear regression) line, , is normally determined as an estimate from a collection of sample data values consisting of values in the scope of the experiment and the corresponding values observed. One common way of estimating the line is the Method of Least Squares. The goal of this method is to create a line that minimizes the summation of the residual error squared. The residual error values are the distances of each sample data point from the resulting best fit line. An example of the graphical representation of residual error is shown below:
Let us use ei to represent each residual error, yi to represent each observed value of y, and ŷi to represent the value of on the estimated line for each yi . The method of least squares involves minimizing Σei = Σ(yi - ŷi)2.This is done using partial derivatives, which yield the following formulas for a(y intercept estimate) and b(slope estimate):
b = ( nΣxiyi - (Σxi)(Σyi) ) / ( nΣxi2 - (Σxi)2 )
a = ( Σyi - bΣxi) / n
The line created using the Method of Least Squares above is characterized by two distinct features:
- Always goes through the point (Xbar, Ybar), where Xbar and Ybar are the average of all sample data xi and yi
- Residual errors are split so that the positive residuals cancel the negative residuals