Simple Linear Regression (Presentation)

You might also like

Download as pptx, pdf, or txt
Download as pptx, pdf, or txt
You are on page 1of 37

1

Simple linear
regression
Presented by Ayesha Ashmal
Areeba kiani
Laiba zahra
Fatima ikram
Rafia nadeem
Table of Contents:
1. What is regression model and what is simple linear
regression. (Ayesha)
2. Assumptions of simple linear regression. (Areeba)
3. How to perform simple linear regression. (Rafia)
4. Interpreting the results. (Fatima)
5. Presenting the results. (Laiba)
What is Regression
Model?
▸ This model proves a function that describes the
relationship between one or more independent
variables and a response, dependent, or target
variables
▸ Example; Age and height can be described using a
linear regression model.
Regression Model
▸ Regression is a statistical technique for investigating and
modelling the relationship between variables.
▸ Applications of regression are numerous and occur in
almost every field including engineering, the physical and
social sciences, and the biological sciences.
▸ Usually, the investigated seeds to ascertain the casual
effects of one variable upon another the effect of price
increase upon demand for example for the effect of
changes in the money supply upon the inflation rate.
What is Simple Linear
Regression?
▸ Statistical tool that gives us the ability to estimate
the mathematical relationship between a
dependent variable usually called why and an
independent variable known as x.
Simple Linear Regression
▸ Mathematical relationship between two variables
X and y is a linear relationship.
▸ Effect relationship, the independent variable is the
cause, and the dependent variable is the effect.
▸ Least square linear regression is a method for
predicting the value of dependent variable why,
based on the value of independent variable x.
▸ Dependent variable; the variable we wish to
explain.
▸ Independant variable; the variable used to explain
the dependent variable.

Assumptions of
1
simple linear
regression
Presented by Areeba kiani
Assumptions in
linear regression
▸ Simple linear regression is a parametric test, meaning
that it makes certain assumptions about the data.
These assumptions are:
▸ Homogeneity of variance (homoscedasticity)
▸ Independence of observations
▸ Normality
▸ Linear regression makes one additional assumption
▸ The relationship between the independent and
dependent variable is linear
12
Homogeneity of
variance
▸ population variances (i.e., the distribution, or “spread,”
of scores around the mean) of two or more samples are
considered equal.
▸ The size of the error in our prediction doesn’t change
significantly across the values of the independent
variable.

13
Independence of
observations
▸ Two observations are independent if the occurrence
of one observation provides no information about the
occurrence of the other observation.
▸ The observations in the dataset were collected using
statistically valid sampling methods, and there are no
hidden relationships among observations.

14
Normality
▸ The data follows a normal distribution.
▸ In a normal distribution, data is symmetrically
distributed with no skew.

15

▸ The relationship between the
independent and dependent
variable is linear: the line of
best fit through the data
points is a straight line (rather
than a curve or some sort of
grouping factor).

16
simple linear regression in Spss

Presented To: Sir Ammar


Presented By: Rafia Nadeem
How to perform a simple linear
regression in Spss:
▸ STEP1:
Step 2:
Step 3:
Step 4:
Step 5:
Step 6:
Interpreting the result.

By fatima ikram
The output table first repeats the formula that was used to generate the
results (‘Call’), then summarizes the model residuals (‘Residuals’), which
give an idea of how well the model fits the real data.
Next is the ‘Coefficients’ table. The first row gives the
estimates of the y-intercept, and the second row gives the
regression coefficient of the model.
Row 1 of the table is labeled (Intercept). This is the y-intercept of
the regression equation, with a value of 0.20. You can plug this into
your regression equation if you want to predict happiness values
across the range of income that you have observed:happiness = 0.20
+ 0.71*income ± 0.018
The next row in the ‘Coefficients’ table is income. This is the row that
describes the estimated effect of income on reported happiness:
The Estimate column is the estimated effect, also called the regression
coefficient or r2 value.
The Std. Error column displays the standard error of the estimate. This
number shows how much variation there is in our estimate of the relationship
between income and happiness
The Pr(>| t |) column shows the p-value. This number tells us how
likely we are to see the estimated effect of income on happiness if the
null hypothesis of no effect were true. Because the p-value is so low (p
< 0.001), we can reject the null hypothesis and conclude that income
has a statistically significant effect on happiness.
The last three lines of the model summary are statistics about the model as
a whole. The most important thing to notice here is the p-value of the
model. Here it is significant (p < 0.001), which means that this model is a
good fit for the observed data.
Presenting the
results and
Limitations
Presented by laiba zahra
Presenting the result
▸ When reporting your results, include the estimated effect (i.e. the regression
coefficient), standard error of the estimate, and the p-value. You should also
interpret your numbers to make it clear to your readers what your regression
coefficient means
▸ It can also be helpful to include a graph with your results. For a simple linear
regression, you can simply plot the observations on the x and y axis and then
include the regression line and regression function

34
▸ For example:
We found a significant relationship (p < 0.001) between
income and happiness (R2 = 0.71 ± 0.018), with a 0.71-


unit increase in reported happiness for every $10,000
increase in income.

35
Limitations of simple
linear regression
▸ Main limitation of Simple Linear Regression is
the assumption of linearity between the dependent
variable and the independent variables. In the real
world, the data is rarely linearly separable. It
assumes that there is a straight-line relationship
between the dependent and independent variables
which is incorrect many times

36

37

You might also like