Download as pdf or txt
Download as pdf or txt
You are on page 1of 11

Second Course in Statistics Regression

Analysis 7th Edition Mendenhall


Solutions Manual
Visit to download the full and correct content document: https://testbankdeal.com/dow
nload/second-course-in-statistics-regression-analysis-7th-edition-mendenhall-solution
s-manual/
Chapter 7 7-1

Chapter

Some Regression Pitfalls 7


7.1 Model is less reliable.

7.2 a. Multicollinearity first increases the likelihood of rounding errors in our calculations.
Second, the results themselves may not seem to make sense (i.e., a significant F test for model
utility but no significant t tests among the individual parameter estimates). A third problem is
that a high correlation between variables may affect the sign of the parameter estimates, a
positive slope is obtained when a negative relationship was expected.

b. One way to detect multicollinearity is to test for pairwise correlation of the independent
variables. Another way to detect multicollinearity is to look for confusing results in the
regression analysis.

c. When multicollinearity is present, one way of dealing with the problem is to remove all but one
of the highly correlated variables, possibly by using stepwise regression. When using the model
for estimation and prediction only, it is not necessary to drop any independent variables. It is
necessary, though, to make sure the values of the x variables fall within the ranges of x used in
the experiment. The presence of multicollinearity when making inferences about the β 's in the
model is a dangerous problem. The solution is to use a designed experiment to break up the
patterns of multicollinearity. To reduce rounding errors in polynomial models, it is useful to
code the x values so that the correlation between the subsequent powers of x is reduced. Another
way of reducing rounding errors due to multicollinearity is to use ridge regression which
provides biased estimates of the β 's in the model, but these estimates have smaller standard
errors than their least squares analogs.

7.3 Another transformation that might work in Example 7.4 would be to use x = p 2 ;ln( p); e-p .

7.4 a. Even though r = .983 is rather large, you must take into caution about deducing a cause-and-
effect between the number of females in managerial positions and the number of females with a
college degree. The independent variables could not be controlled, but there could be a strong
relationship between these two variables.

b. There could be a problem with multicollinearity.

7.5 a. Based on the correlation matrix, there is no evidence of extreme multicollinearity. None of the
pairwise correlations are larger than .5 in absolute value. Thus, it does not appear that
multicollinearity is a big problem in this exercise.

b. No, there is no evidence of extreme multicollinearity. The global test is significant


( F = 32.47 with p − value < .001) indicating that at least one of the variables contributes to
the prediction of urban/rural rating. However, 3 of the 6 independent variables, when tested for
significance in predicting urban/rural rating, are significant. Multicollinearity is present if the
global test is significant, but none (or very few) of the individual variable tests are significant.

Copyright © 2012 Pearson Education, Inc. Publishing as Prentice Hall.


7-2 Some Regression Pitfalls

7.6 The "experimental region", which is defined in the text as the range of values of the independent
variables in the sample data. Ultimately, we do not want to predict the DV for values of the IVs
outside the experimental region (the extrapolation problem).

Since there are two IV's (depth and ice type), we need to find the "joint" range. Using Minitab, we get
the output below:

Descriptive Statistics: depth

Variable icetype N Mean StDev Minimum Maximum

depth First-Year 88 0.13080 0.08026 0.02000 0.36000


Landfast 196 0.31160 0.1920 0.00000 0.86000
Multi-Year 220 0.27527 0.12646 0.07000 0.64000

Note that the range for the First-Year ice is Range = .36 − .02 = .34, the range for the Landfast ice is
Range = .86 − .00 = .86, and the range for the Multi-Year ice is Range = .64 − .07 = .57.

Avoid selecting values of depth outside these ranges to avoid extrapolation errors.

7.7 a. There is not evidence of extreme multicollinearity since aggressive behavior is highly correlated
with all three independent variables.

b. Since narcissism is only somewhat correlated with irritability and trait anger, these two variables
should not cause much concern about multicollinearity. However, if you try to use all three
independent variables which would include aggressive behavior a multicollinearity problem
could occur.

7.8 Based on these results, we would say that multicollinearity is a severe problem in this analysis. The
model predicts the amount of carbohydrate solubilized during steam processing of peat (y) very well.
We disagree with the bioengineer's statement. Note that since the given correlations between pairs of
independent variables of the model in this problem are large, these indicate the presence of
multicollinearity (see “Detecting Multicollinearity in the Regression Model”, page 365).

7.9 When fitting the pth-order polynomial regression model, two requirements must be met:
1. The number of levels of x must be greater than or equal to ( p+1). For a second-order model
( p = 2), we must have at least p + 1 = 3 levels of x.
2. The sample size n must be greater than ( p + 1) in order to allow sufficient degrees of freedom
for estimating σ 2 . For a second-order model, we must have n greater than ( p +1) = 3.

Note that for our sample data, requirement #1 is satisfied since x = 1, 2, or 5. However, requirement
#2 is not satisfied since n = 3. If we attempt to fit this model, we will have
n −( k + 1) = n − 3 = 0 df for estimating σ 2 and, thus, will be unable to test model adequacy.

Copyright © 2012 Pearson Education, Inc. Publishing as Prentice Hall.


Chapter 7 7-3

7.10 a. n=9 SS xx = 17200 βˆ1 = .62597


∑ x = 3720 SS yy = 7222.22 βˆ0 = 5.7106
∑ y = 2380 SS xy = 10766.67 SSE = 482.62
2
∑ x = 1554800 s 2 = 68.95 s = 8.30
2
∑ y = 636600 x = 413.33
∑ xy = 994500 y = 264.44

The least squares line is yˆ = 5.7106 + .62597x.

1 ( xp − x )
2
b. A 95% prediction interval for y is yˆ ± tα /2 s 1 + +
n SS xx

where t.025 = 2.365 is based on n − 2 = 7 df.

For x p = 300, yˆ = 5.7106 + .62597 (300) = 193.50. Substituting into the formula yields:

1 (300 − 413.33)2
193.50 ± ( 2.365)(8.30) 1 + + ⇒ 193.50 ± 26.76 or (166.74, 220.26)
9 17, 200

c. Note that the values of x in the sample range from 340 to 480 pounds. The prediction interval
should be used cautiously, or not at all, since the value xp = 300 lies outside the experimental
region.

7.11 a. Using SAS to fit the model, we get:

Model: MODEL1
Dependent Variable: Y

Analysis of Variance

Sum of Mean
Source DF Squares Square F Value Prob>F

Model 1 494.28131 494.28131 253.370 0.0001


Error 23 44.86909 1.95083
C Total 24 539.15040

Root MSE 1.39672 R-square 0.9168


Dep Mean 12.52800 Adj R-sq 0.9132
C.V. 11.14880

Parameter Estimates

Parameter Standard T for H0:


Variable DF Estimate Error Parameter=0 Prob > |T|

INTERCEP 1 2.743278 0.67520594 4.063 0.0005


X1 1 0.800976 0.05032017 15.918 0.0001

Copyright © 2012 Pearson Education, Inc. Publishing as Prentice Hall.


7-4 Some Regression Pitfalls

The least squares equation is yˆ = 2.74 + .80 x1.


Yes, t = 15.92. The p-value associated with the tar content ( x1 ) variable is p= .0001. For
α > .0001, tar content is a useful predictor of carbon monoxide content.

b. Using SAS to fit the model, we get:

Model: MODEL1
Dependent Variable: Y

Analysis of Variance

Sum of Mean
Source DF Squares Square F Value Prob>F

Model 1 462.25591 462.25591 138.266 0.0001


Error 23 76.89449 3.34324
C Total 24 539.15040

Root MSE 1.82845 R-square 0.8574


Dep Mean 12.52800 Adj R-sq 0.8512
C.V. 14.59493

Parameter Estimates

Parameter Standard T for H0:


Variable DF Estimate Error Parameter=0 Prob > |T|

INTERCEP 1 1.664666 0.99360176 1.675 0.1074


X2 1 12.395406 1.05415188 11.759 0.0001

The least squares equation is yˆ = 1.66 + 12.40x2 .

Yes, t = 11.76. The p-value associated with the nicotine content ( x2 ) variable is p= .0001. For
α > .0001, nicotine content is a useful predictor of carbon monoxide content.

c. Using SAS to fit the model, we get:

Model: MODEL1
Dependent Variable: Y
Analysis of Variance

Sum of Mean
Source DF Squares Square F Value Prob>F
Model 1 116.05651 116.05651 6.309 0.0195
Error 23 423.09389 18.39539
C Total 24 539.15040
Root MSE 4.28898 R-square 0.2153
Dep Mean 12.52800 Adj R-sq 0.1811
C.V. 34.23519
Parameter Estimates

Parameter Standard T for H0:


Variable DF Estimate Error Parameter=0 Prob >|T|
INTERCEP 1 -11.795271 9.72162623 -1.213 0.2373
X3 1 25.068198 9.98028209 2.512 0.0195

Copyright © 2012 Pearson Education, Inc. Publishing as Prentice Hall.


Chapter 7 7-5

The least squares equation is yˆ = −11.80 + 25.07 x3 .


Yes, t = 2.51. The p-value associated with the weight ( x3 ) variable is p= .0195. For
α > .0195, weight is a useful predictor of carbon monoxide content.

d. We see the estimate signs have changed dramatically when compared to the model with all three
variables present.

7.12 If there is no severe correlation between x1 and x2 , then the economist could try to fit the model.

7.13 a. There may be multicollinearity between the relative error in estimating effort and the company
role of estimator which could make the sign of βˆ in the model have an opposite sign than what
1
is expected.

b. If there is no data collected for project leaders with less than 20% accuracy this forces all values
of x8 to be x8 = 1. Thus, all the β 3 is not estimable.

7.14 If Nickel is highly correlated with each of the other ten potential independent variables, then
multicollinearity could be a problem. You could drop Nickel altogether, or transform (code) this
variable to reduce the multicollinearity.

7.15 a. No multicollinearity since we use all three levels variables, use DOTest and LBERATIO.

b. Using SAS, the fitted full interaction model results are:

The REG Procedure


Model: MODEL1
Dependent Variable: PRICE

Analysis of Variance

Sum of Mean
Source DF Squares Square F Value Pr > F

Model 6 689573044 114928841 1694.46 <.0001


Error 210 14243534 67826
Corrected Total 216 703816578

Root MSE 260.43493 R-Square 0.9798


Dependent Mean 1126.66359 Adj R-Sq 0.9792
Coeff Var 23.11559

Parameter Estimates

Parameter Standard
Variable DF Estimate Error t Value Pr > |t|

Intercept 1 -9.24852 40.30694 -0.23 0.8187


DOTEST 1 0.86618 0.03554 24.37 <.0001
STATUS 1 5.32761 63.97853 0.08 0.9337
DAYS 1 0.14195 0.18918 0.75 0.4539
DOT_STAT 1 0.20858 0.07604 2.74 0.0066
DOT_DAY 1 0.00004264 0.00005855 0.73 0.4673
STAT_DAY 1 -0.04285 0.48301 -0.09 0.9294

Copyright © 2012 Pearson Education, Inc. Publishing as Prentice Hall.


7-6 Some Regression Pitfalls

7.16 For the quadratic (second-order) model, E ( y ) = β 0 + β1 x + β 2 x 2 , we have p = k = 2, where p is the


order of the model and k is the number of β 's in the model excluding β 0 . To fit the model, we
require:

1. At least p + 1 = 3 different values of x in order to estimate all three β 's.

2. Sample size n > ( k + 1) = 3 in order to have sufficient degrees of freedom for estimating σ 2 .

7.17 Two levels each; n > 4.

7.18 For the complete second-order model,

E ( y ) = β 0 + β1 x1 + β 2 x2 + β 3 x1 x2 + β 4 x12 + β 5 x22

we have p = 2 and k = 5 where p is the order of the model and k is the number of β 's in the model
excluding β 0 . To fit the model, we require:

1. At least p + 1 = 3 different values of x1 and at least 3 different values of x2 in order to estimate


all five β 's.

2. Sample size n > ( k + 1) = 6 in order to have sufficient degrees of freedom for estimating σ 2 .

7.19 There are signs of multicollinearity since the VIF values for INLET-TEMP, AIRFLOW, and POWER
are greater than 10. In order to modify the model you could drop either POWER or AIRFLOW since
their VIF values are very high and redo the analysis. Or you could transform (code) the highly
correlated variables to reduce the multicollinearity. Include only one of the three variables.

7.20 a. There appears to be a curvilinear trend.

Copyright © 2012 Pearson Education, Inc. Publishing as Prentice Hall.


Chapter 7 7-7

b.

x y ln x ln y
54 6 3.989 1.792
42 16 3.738 2.773
28 33 3.332 3.497
38 18 3.678 2.890
25 41 3.219 3.714
70 3 4.248 1.099
48 10 3.871 2.303
41 14 3.714 2.639
20 45 2.996 3.807
52 9 3.951 2.197
65 5 4.174 1.609

c. Let y* = ln y and x* = ln x. Preliminary calculations follow:

n = 11 SS x* x* = 1.5675 βˆ1 = −2.16985


∑ x * = 40.87 SS y* y* = 7.7475 βˆ0 = 10.6364
∑ y * = 28.32 SS x* y* = −3.4011 SSE = .3675
2
∑ ( x*) = 153.42 s 2 = .0408
2
∑ ( y*) = 80.65 x * = 3.715 s = .2021
∑ x * y * = 101.81 y * = 2.574

Copyright © 2012 Pearson Education, Inc. Publishing as Prentice Hall.


7-8 Some Regression Pitfalls

The least squares prediction equation is yˆ * = 10.6364 - 2.16985x *

or = 10.6364-2.16985 ln ( x )

H 0 : β1 = 0
H a : β1 ≠ 0
βˆ1 −2.16985
Test statistic: t = = = −13.44
s / SS xx .2021 / 1.5675

Rejection region: α = .05, n − 2 = 9 df, t.025 = 2.262


Reject H 0 if t > 2.262 or t < −2.262

Conclusion: Reject H 0 at α = .05. There is sufficient evidence to indicate that the transformed
model is adequate for predicting y.

d. When x = 30, lnx = ln 30 = 3.401 and = 10.6364-2.16985(3.401) = 3.256

To obtain the predicted value, , we use the inverse transformation (antilog):

ŷ = e = e3.256 = 25.95

7.21 a. Some preliminary calculations are:


∑ x1 = 450.1 ∑ y = 1800 ∑ x1 y = 54013.5
2 2
∑ x1 = 13906.27 ∑ y = 216900.86
∑ x1 ∑ y 450.1(1800)
SS x1 y = ∑ x1 y − = 54013.5 − = 1.5
n 15
2
2 (∑ x1 ) 450.12
SS x1x1 = ∑ x1 − = 13906.27 − = 400.26933
n 15
2
(∑ y ) 18002
SS yy = ∑ y 2 − = 216900.86 − = 900.86
n 15

The correlation coefficient is:


SS x1 y 1.5
r= = = .0025
SS x1x1 SS yy 400.26933(900.86)

Because the value of r is so small, there is no evidence of a linear relationship between x1 and y.

b. Some preliminary calculations are:


∑ x2 = 1049.8 ∑ y = 1800 ∑ x2 y = 126844.85
2 2
∑ x2 = 77911.98 ∑ y =216900.86
∑ x2 ∑ y 1049.8(1800)
SS x2 y = ∑ x2 y − = 126844.85 − = 868.86
n 15

Copyright © 2012 Pearson Education, Inc. Publishing as Prentice Hall.


Chapter 7 7-9

2
(∑ x2 ) 1049.82
SS x2 x2 = ∑ x22 − = 77911.98 − = 4439.97733
n 15
2
(∑ y ) 18002
SS yy = ∑ y 2 − = 216900.86 − = 900.86
n 15

The correlation coefficient is:


SS x2 y 868.86
r= = = .434
SS x2 x2 SS yy 4439.97733(900.86)

Since the value of r is not very close to 1, there is little evidence of a linear relationship between
x2 and y.

c. Based on the values of the correlation coefficients computed in parts (a) and (b), there is no
evidence that the model will be useful for predicting the sale price.

d. Using SAS, the output is:

DEP VARIABLE: Y
ANALYSIS OF VARIANCE

SOURCE DF SUM OF SQUARES MEAN SQUARE F VALUE PROB > F


MODEL 2 900.72221 450.36111 39222.343 0.0001
ERROR 12 0.13778711 0.01148226
C TOTAL 14 900.86000

ROOT MSE 0.1071553 R-SQUARE 0.9998


DEP MEAN 120 ADJ R-SQ 0.9998
C.V. 0.08929609

PARAMETER ESTIMATES

PARAMETER STANDARD T FOR HO:


VARIABLE DF ESTIMATE ERROR PARAMETER=0

INTERCEP 1 !45.15413641 0.61141807 !73.851


X1 1 3.09700789 0.01227443 252.314
X2 1 1.03185903 0.003684173 280.079

VARIABLE DF PROB > |T|

INTERCEP 1 0.0001
X1 1 0.0001
X2 1 0.0001

From the printout, the least squares line is yˆ = −45.154 + 3.097x1 + 1.032x2 .

To determine if the model is adequate, we test:


H 0 : β1 = β 2 = 0
H a : At least one β i ≠ 0

The test statistic is F = 39222.34

Copyright © 2012 Pearson Education, Inc. Publishing as Prentice Hall.


7-10 Some Regression Pitfalls

The p-value is .0001. Since this p-value is so small, H o is rejected. There is sufficient evidence
to indicate the model is adequate for predicting sale price.

The value of R 2 is .9998. This indicates that 99.98% of the sample variation in the sale prices is
explained by the model including x1 and x2 .

These results disagree with those in part (c).

e. Some preliminary calculations are:


2
∑ x1 = 450.1 ∑ x1 = 13906.27 ∑ x1 x2 = 30304.09
2
∑ x2 = 1049.8 ∑ x2 = 77911.98
∑ x1 ∑ x2 450.1(1049.8)
SS x1x2 = ∑ x1 x2 − = 30304.09 − = −1196.90867
n 15
2
(∑ x1 ) 450.12
SS x1x1 = ∑ x12 − = 13906.27 − = 400.26933
n 15
2
2 ( ∑ x2 ) 1049.82
SS x2 x2 = ∑ x2 − = 77911.98 − = 4439.97733
n 15

The correlation coefficient is:


SS x1x2 −1196.90867
r= = = −.8998
SS x1x2 SS x2 x2 400.26933(4439.97733)

Because the correlation coefficient between x1 and x2 is very close to −1, it implies that x1 and
x2 are highly correlated.

f. In this case, we would not want to throw out a redundant variable. The models with just one
independent variable are not significant. However, the model with both independent variables,
even though they are highly correlated, is very significant.

7.22 a. Variables that are moderately or highly correlated have correlation coefficients of .5 or higher in
absolute value. There are only two pairs of variables which are moderately or highly correlated.
They are "year GRE taken" and "years in graduate program"
(r = −.602), and "race" and "foreign status" (r = −.515).

b. When independent variables that are highly correlated with each other are included in a
regression model, the results may be confusing. Highly correlated independent variables
contribute overlapping information in the prediction of the dependent variable. The overall
global test can indicate that the model is useful in predicting the dependent variable, while the
individual t-tests on the independent variables can indicate that none of the independent variables
are significant. This happens because the individual t-tests tests for the significance of an
independent variable after the other independent variables are taken into account. Usually, only
one of the independent variables that are highly correlated with each other are included in the
regression model.

7.23 Since df ( Error ) = 0 , the researcher's sample size is not sufficient to estimate s 2 . Thus, no estimate of
the variation in the y values exists and no test of the model adequacy can be made.

Copyright © 2012 Pearson Education, Inc. Publishing as Prentice Hall.

You might also like