Professional Documents
Culture Documents
This Content Downloaded From 157.40.102.185 On Sat, 02 Oct 2021 19:55:43 UTC
This Content Downloaded From 157.40.102.185 On Sat, 02 Oct 2021 19:55:43 UTC
This Content Downloaded From 157.40.102.185 On Sat, 02 Oct 2021 19:55:43 UTC
REFERENCES
Linked references are available on JSTOR for this article:
https://www.jstor.org/stable/2288345?seq=1&cid=pdf-
reference#references_tab_contents
You may need to log in to JSTOR to access the linked references.
JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide
range of content in a trusted digital archive. We use information technology and tools to increase productivity and
facilitate new forms of scholarship. For more information about JSTOR, please contact support@jstor.org.
Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at
https://about.jstor.org/terms
Taylor & Francis, Ltd. and American Statistical Association are collaborating with JSTOR to
digitize, preserve and extend access to Journal of the American Statistical Association
The time series regression models in which the errors of regressive moving average (ARMA) model. Assuming
regression equations follow stationary or nonstationary that Zt is an autoregressive process with finite order, Dur-
autoregressive moving average models are considered. bin (1960) proposed a two-stage procedure that yields
Convergence properties of the sample autocorrelation asymptotically efficient estimates in linear models, and
function of observed series and the least squares esti- Gallant and Goebel (1976) provided a procedure for es-
mates of the linear regression parameters are shown. timating the unknown parameter 3 in nonlinear regres-
Based upon these results, a procedure for specifying the sion settings. Fuller (1976) discussed some properties of
tentative order of the mixed ARMA errors is proposed. model (1.1) in Chapter 9. More recently, using the Kal-
Two examples are given. man filter techniques, Harvey and Phillips (1979) consid-
ered the maximum likelihood estimation under the same
KEY WORDS: Autoregressive moving average model;
conditions as those of Pierce (1971a).
Extended sample autocorrelation function; Intervention;
In this article we also consider the model (l.1) but allow
Regression; Time series.
the time process Zt to be nonstationary. Such nonsta-
1. INTRODUCTION tionary regression models are of interest because, first of
all, they provide a simple method for removing the effects
A useful and extensive class of models for describing of certain outlying observations in modeling nonstation-
economic, business, and environmental data is the col- ary time series data; for example, see Chang (1982) and
lection of time series regression models. It consists of Hillmer, Bell, and Tiao (1982). Second, these models are
models in the form useful in studying the so-called "trading day" effects in
analyzing monthly economic data. For instance, the
Yt = flX,, ) + Z. (1.1)
model
where Y, is the series of interest, Zt is a time series
7
proc-
ess that is unobservable, X, denotes a set of input (or
Yt= iXit + Zt
exogenous) variables, , is a vector of parameters, and i = I
under the assumption that the time series component Zt some asymptotic properties of the observed series Yt
is stationary. Anderson (1954) gave a review of earlier when Zt is nonstationary. Section 2 states the model used,
work on the regression analysis when autocorrelation ex-the assumptions imposed, and the main results. The cor-
ists. Hannan (1971) proved consistency properties of theresponding proofs are given in the Appendix. Based upon
weighted least squares (LS) estimates of the parameter these results, Section 3 describes the proposed order se-
,1 when fiX,, 1) is nonlinear and Z, has a continuous lection procedure for the linear regression time series
spectrum function. Pierce (1971a) considered the LS es- models, and Section 4 applies the procedure to two ex-
amples. Finally, Section 5 addresses some nonlinear sit-
timation when the model is linear and Zt follows an auto-
uations, and Section 6 gives a brief discussion.
118
model (2.7) with m 1, The above corollary can be generalized to obtain con-
n sistent LS estimates for the nonstationary coefficients
(a) E (Z, - ) = 2 0(n2m); (2.8) Ui's in U(B) if 81 c 1. This can also be proved by the
same methods as those in Tiao and Tsay (1983) and can
n
be used to spot the nonstationary factor U(B) of Zt in
(b) (Z t Z)2J = O (n2m); (2.9) practical modeling.
Next consider the situation that some 8j in (2.11) is
n greater than the leading order 2m of Z,; that is, consider
(C) (Zt - Z)(Wt+h - W) = Op(n m +) (2.10) the case where some Xi, is dominating. In this case, the
linear regression
residuals of previous regressions as newly added regres- The motivation behind this procedure is as follows: (i)
sors to accommodate the effects of the MA part on the under the condition l(b), the true time series Z, is non-
AR estimates. As the iteration advances, the MA effects stationary with the highest multiplicity of nonstationary
will reduce and eventually vanish when the number of roots being greater than the orders of Xi, and by Theorem
iterations exceeds the true MA order q. A recursive al- 2.1, the SACF would provide consistent estimates for
gorithm for computing the iterated estimates of AR pa- some nonstationary factors of Z,; and (ii) under the sit-
rameters has also been developed. uation 1(c), the time series component is no longer dom-
A useful property of the ESACF for model identifi- inating and removing the effects of X, becomes necessary.
cation is that the pth ESACF of an ARMA(p, q) model,
stationary or not, has exactly the same asymptotic "cut- 4. EXAMPLES
ting-off' property as that of the ordinary SACF of an
MA(q) model. In applications, one can simply search Two examples are presented here for the proposed
from the ESACF table for this cutting-off feature to spec- modeling procedure and for those properties shown in
ify the order (p, q) of a time series model. The readers Section 2.
are referred to Tsay and Tiao (1984) for details and ex-
Example 1. Consider a linear regression model with
amples. Here we employ this ESACF approach to iden-
nonstationary time series errors. Two hundred observa-
tify the time series model after the effects of input vari-
tions were generated from the model
ables (i.e. f(X,, 0)) have been approximately removed.
The asymptotic properties stated in Section 2 provide Yt= 5.0 + 2.OX, + Zt, (4.1a)
some clues for accessing the nonstationary factors of Zt
(1 - B)(1 - .5B)Zt = (1 + .5B)a,, (4.lb)
and the effects of X,. For example, one can employ the
where Xt and a, are iid N(5, 16) and N(O, 1), resp
SACF of Y, or the LS estimates of stepwise AR fittings
to specify the nonstationary factors of Z, if E = The data are plotted in Figure 1 and the fitted linear
0(n) for all i. On the other hand, one would fit a linear regression equation, assuming that Zt's of (4.1a) were
regression to remove the input effects when information independent, is
indicates that Z, is stationary. In general, if the station-
arity of Zt is uncertain, we recommend the following pro-
Yt= -.104 + 2.43Xt + e,, (4.2)
cedure for the model specification.
(1.54) (.24)
Step 1. Compute the SACF of Y, and that of the esti-
where the values in parentheses are the corresponding
mated residuals et of the regression (2.14). (a) If the SACF
standard deviations. The biases of the estimates are
of et decays rapidly, go to Step 2 and treat et as Zt, an
clearly shown in (4.2).
estimate for the time series component Zt. (b) If both
The SACF's of the original Yt and the residuals e, of
SACF's fail to die out quickly but satisfy the same dif-
(4.2) are given, respectively, in (a) and (b) of Table 1,
ference equation, transform Y, and Xi,'s according to the
which clearly indicates that a nonstationary factor (1 -
common difference equation and return to Step 1 with
B) exists in the time series process. Following the pro-
these transformed series as observed values. (c) If both
posed procedure, we take first difference on both Y, and
SACF's fail to die out quickly and satisfy two difference
X, and iterate the above step. Table 1(c) gives the SACF
equations, remove the effects of those input variables Xit
of the residuals E, obtained from the regression,
that possess the highest order, and go to Step 1 with the
adjusted Yt and the remaining input variables.
Step 2. Compute the ESACF table of Z, to identify the VY, = 1.99 VXt + A,, V = (1 -B), (4.3)
order (p, q) of a time series model. (.02)
Lag SACF
a. The SACF of Yt
1-12 .74 .74 .77 .72 .66 .66 .63 .59 .61 .54 .51 .49
13-24 .47 .44 .42 .39 .37 .35 .32 .29 .26 .23 .18 .18
b. The SACF of et
1-12 .97 .95 .92 .89 .86 .83 .79 .76 .73 .68 .66 .62
13-24 .59 .56 .53 .50 .46 .43 .40 .37 .35 .32 .29 .26
c. The SACF of i
1-12 .73 .40 .22 .14 .09 .06 .04 .03 .02 - .04 - .08 -.05
13-24 -.07 -.05 -.00 .02 .01 -.03 -.05 -.05 -.05 -.07 -.08 -.08
MA
AR 0 1 2 3 4 5 6
X denotes the absglute value of the corresponding entry as greater than or equal to Yt= rXtl + 02x2t + Zt,
2n- 5, and 0 otherwise.
and this time the SACF fails to support any nonstation- where Yt is the traffic volume, Xl, = E,, and X21 = Et
arity behavior of the differenced series. The ESACF table (T - 1962) with T representing the year and et the pro
and its simplified version for E, in (4.3) are then computed
portion of the new year period in the tth month. In th
process of arriving at this model, Liu noticed that the first
and presented in Table 2. An ARMA(1, 1) model is clearly
suggested for the time series component from this table. difference (1 - B) of Z, is clearly suggested by the SAC
Thus, as expected, the generating model (4.1) is correctly of Y,, but the seasonal difference (1 - B'2) is not by th
specified by the proposed procedure. SACF of (1 - B) Yt. Moreover, the need of a seasonal
For this example, a brief discussion is in order. In Tabledifference is indicated by the SACF of (1 - B)V, where
Vt is an adjusted series of Yt by removing the calendar
1, the SACF of e, seems to outperform that of Y, in pin-
pointing the nonstationary factor (1 - B). However, it is
effects, that is, the Chinese New Year effects.
incorrect in this situation to identify a model for the time Assuming that model (4.4) is adequate in describing the
series component basing upon the differenced process (1 data, we now give some theoretical explanations for these
- B)e,, a procedure likely to be employed in practice. phenomena observed by Liu. Since 1 is a double root of
The reason is as follows. The input variable X, in this the nonstationary AR part of Zt, by a result in Tiao and
Lag SACF
1-12 -.15 .07 .10 .06 -.04 .08 .02 -.09 .13 -.10 -.02 -.01
13-24 -.00 -.00 -.01 .02 -.02 -.01 -.07 -.01 .02 .02 -.06 .04
5. NONLINEAR REGRESSION MODELS model is different from that in a stationary AR(1) model;
see, for example, Dickey and Fuller (1979) and Ahtola
We consider some nonlinear situations; that is, f(X,,
(1983).
,) is nonlinear. In these cases, Assumption 5 of Section
2. Regression Specification. In this article, we assume
2 is no longer valid, and must be replaced by one that
that the functional form of the regression equation is
assures the existence of a unique solution for the problem
known. The question of selecting the best regression
encountered. Moreover, unlike that of linear cases, the
model, however, is likely to arise in applications. One
impacts of exogenous variables now also depend on the
possible solution is imbedded in the diagnostic checkings.
true structure of flX,, ,). For example, the order of thespecifically, one incorporates at the beginning stage
More
input variable t itself in (2.4) is 0(n3) but that of cos(tP14)
of a modeling process all the available input variables that
and sin(t134) is only 0(n). The SACF of Y, would then
are known or suspected to have effects on the response
satisfy a homogeneous difference equation if the time se-
series Yt and then one deletes those unimportant varia-
ries component Zt is nonstationary and the number of
bles, judging in terms of the corresponding standard de-
observations is large, because any nonstationary root
viations, towards the end. Of course, some approaches
would give Z, with order 0(n2) that suffices to override
that can identify models for both the regression and the
the effects of exogenous variables. On the other hand,
time series components at the model specification stage
the effect of the input variable t in (2.5) always upsets
of an analysis need to be developed.
that of any nonstationary factors in Z, if 12 > 0. Mainly
3. Differencing. Finally, it may be worthwhile men-
because nvI(lE exp(2Q2t)) O' 0 as n -*oo for any finite vtioning that differencing the oberved nonstationary series
if 12 > 0. Thus for (2.5), nonlinear regression must be
Y, to achieve stationarity in the time series component
fitted to remove the input effects before attempting to
Z, may also reduce the effects of exogenous variables.
specify the order of Z,. In general, similar results such For instance, the first difference (1 - B) removes the
as Theorems 2.1 and 2.2 of linear cases can be established
constant term from flX,, ,) and transforms the linear
whenever the orders of the effects of input variables are
trend, if any, into a constant.
accessible. The modeling procedure of Section 3 can then
be extended to the nonlinear situation. In practice, if the APPENDIX
orders of input effects are unavailable, such as the sign
of 12 in (2.5) is unknown a priori, we recommend that Proof of Theorem 2.1
one fit the nonlinear regression first and then choose a Since Y = 'I ,Xi + Z, Yt - Y = = j3B(Xi-
proper procedure based upon the fitted results. Xi) + (Zt - Z). By this relationship, the assump
> bi, Cauchy-Schwarz inequality (2.12), and properties
6. CONCLUDING REMARKS (2.8) and (2.9), it is readily shown that
n
In this article, we have proved some properties for the
regression models with nonstationary time series errors. E (Yt - 1)2 = Op(n2m) (A.1)
These results provide information on when and how to
remove the effects of exogenous variables in the process and
1. Simultaneous Estimation. After a model is specified Letting Vt = UI(B)Y, and W, = UI(B)Zt, we have that
for the time series component Z,, all the parameters in k
model (2.3) should be estimated simultaneously. This can Vt= i[U,(B)Xit] + Wt (A.3)
be done by using either the nonlinear least squares or the and i=
maximum likelihood methods. In case of stationary
ARMA errors, a useful state space procedure for eval- U1 (B)ry(l)
uating the likelihood function was proposed by Harvey n ~ I~ n
and Phillips (1979). Two problems, however, remain open _ t(, )a E (yt
1 I~~~~+d + 1
when Z, is nonstationary. First, the initial value K used
by Harvey and Phillips might become critical because
- Y)(V - V)] (A.4)
nonstationary series tend to have long memory. It is then
necessary to have large sample sizes to reduce the effect where denotes asymptotic equivalence. Since the high-
of initial values. Second, the distribution properties such est multiplicity of the nonstationary AR roots of W, is m
as the asymptotic normality of the estimates become un- - 1 and 2m > 51, it is clear from (2.8) and (A.3) that
known. For instance, it has been shown that the distri-
bution of the LS estimate (asymptotically equivalent to
the MLE) of the AR parameter in a nonstationary AR(l)
E ( V, - V)2 = Op(n8) (A.5)
I+ di + 1