New Frontier
New Frontier
SUMMARY
In the 20 years following the publication of the ARCH model, there has been a vast quantity of research
uncovering the properties of competing volatility models. Wide-ranging applications to financial data have
discovered important stylized facts and illustrated both the strengths and weaknesses of the models. There are
now many surveys of this literature. This paper looks forward to identify promising areas of new research.
The paper lists five new frontiers. It briefly discusses three—high-frequency volatility models, large-scale
multivariate ARCH models, and derivatives pricing models. Two further frontiers are examined in more
detail—application of ARCH models to the broad class of non-negative processes, and use of Least Squares
Monte Carlo to examine non-linear properties of any model that can be simulated. Using this methodology, the
paper analyses more general types of ARCH models, stochastic volatility models, long-memory models and
breaking volatility models. The volatility of volatility is defined, estimated and compared with option-implied
volatilities. Copyright 2002 John Wiley & Sons, Ltd.
1. INTRODUCTION
Who could imagine 20 years ago, the flowering of research and applications that would develop
around the ARCH model? It certainly was not an instant success. Years went by before anyone
except my students and I wrote a paper on ARCH. But as applications shifted to financial markets,
and as richer classes of models were developed, researchers saw how volatility models could be
used to investigate the fundamental questions in finance. How are assets priced and what is the
tradeoff between risk and return? ARCH models offered new tools for measuring risk, and its
impact on return. They also provided new tools for pricing and hedging non-linear assets such as
options.
This conference and this paper are designed to reflect on these developments and look forward to
the next important areas of research. In this paper I will give a rather brief idiosyncratic assessment
of the important accomplishments of the last 20 years in volatility modelling. Then I will point to
five frontiers on which I think new developments can be expected in the next few years. For two
of these, I will give some new results to show the directions I see developing.
Ł Correspondence to: Robert Engle, Department of Finance, New York University, 44 West 4H, NY, NY 10012, USA.
E-mail: rengle@stern.nyu.edu
Copyright 2002 John Wiley & Sons, Ltd. Received 13 December 2001
Revised 7 June 2002
426 R. ENGLE
soup of volatility models continually amazes. The most influential models were the first: the
GARCH model of Bollerslev (1986), and the EGARCH of Nelson (1991). Asymmetric models
of Glosten, Jaganathan and Runkle (1993) Rabemananjara and Zakoian (1993), Engle and Ng
(1993) and power models such as Higgins and Bera (1992), Engle and Bollerslev (1986), and
Ding, Granger and Engle (1993) joined models such as SWARCH, STARCH, QARCH and many
more. The linguistic culmination might be that of Figlewski (conference presentation, UCSD,
1995), the YAARCH model—an acronym for Yet Another ARCH model. Coupled with these
models was a sophisticated analysis of the stochastic process of data generated by such models as
well as estimators of the unknown parameters. Theorems for the autocorrelations, moments and
stationarity and ergodicity of these processes have been developed for many of the important cases;
see, for example, Nelson (1990), He and Terasvirta (1999a,b) and Ling and McAleer (2002a,b).
Work continues and new models are continually under development, but this is a well-studied
frontier. The limiting distribution of the MLE for GARCH models waited for Lumsdaine (1996)
and Lee and Hansen (1994) for rigorous treatments. There are now a collection of survey articles
that give a good appreciation of the scope of the research. See, for example, Bollerslev, Chou
and Kroner (1992), Bollerslev, Engle and Nelson (1994), Bera and Higgins (1993), and recent
pedagogical articles by Engle (2001) and Engle and Patton (2001). A very recent survey is Li,
Ling and McAleer (2002).
Another topic for ARCH models is their usefulness in trading options. It was initially supposed
that volatility models could give indications of mispricing in options markets leading to trading
opportunities. Early studies such as Engle, Kane and Noh (1994) suggested the profitability of
such strategies. More recent data fails to find evidence of significant trading opportunities, at
least in the US index options market. This is not surprising since GARCH models have a limited
information set and are available to all traders today. The same question is often asked in terms
of forecast accuracy. Do GARCH models out-forecast implied volatility models? The answer is
complex depending upon the statistical approach to forecast evaluation, but generally it is found
that implied volatilities are more accurate forecasts of future volatility than are GARCH models.
See, for example, Poon and Granger (2002).
The theory of asset pricing is based upon the reward for bearing risk. ARCH models have been
developed to measure the price of risk. The first such model was the univariate ARCH-M model
of Engle, Lilien and Robins (1987). Estimation of the CAPM began with Bollerslev, Engle and
Wooldridge (1988) and has been extended and improved by a series of interesting papers including
McCurdy and Stengos (1992), Engel et al. (1995), and de Santis, Gerard and Hillion (1997).
With the introduction of Value at Risk, a new role for ARCH models emerged. A variety
of studies examined the usefulness of volatility models in computing VaR and comparing these
methods with the exponential smoothing approach favoured by Riskmetrics. See, for example,
Christoffersen and Diebold (2000), Christoffersen, Hahn and Inoue (2001) and Alexander (1998).
GARCH methods proved successful but suffered if errors were assumed to be Gaussian.
These chapters of research on ARCH models are full and may have reached the point of dimin-
ishing returns. However, new directions are always available and these are the main focus of
this paper.
new papers. For two of the areas, I will give some new results suggesting a possible direction for
future research.
The study of volatility models within the day is in its infancy yet is a natural extension of the
daily models examined so widely. Several alternative formulations have been introduced including
Andersen and Bollerslev (1997) and Bollerslev, Cai and Song (2000). Such models focus on the
time of day or ‘diurnal’ effect and have the requirement that they be useful forecasting many days
in the future. These models have regularly spaced observations in calendar time but ultimately
it will be desirable to find models based on irregularly spaced data as this is the inherent limit
of high frequency data. Engle (2000) calls such tick data ‘ultra high frequency’ data and gives
some models which indicate that the arrival rate of trades, spreads and other economic variables
may be important variables for forecasting volatility at this frequency. Such a model could give a
continuous record of instantaneous volatility where events such as trades and quote revisions as
well as time itself, modify the volatility estimate.
Continuous time models are ubiquitous in financial theory and derivative pricing. However most
estimation of these models begins with equally observed prices and focuses on the mean process
possibly with jumps. Continuous time stochastic volatility models, possibly with volatility jumps
are a new class of models with interesting derivative implications.
In addition to these models, there is now increasing interest in using intra-daily data to estimate
better daily models. Andersen et al. (2001), for example, build models based upon ‘realized
volatility’ and Andersen and Bollerslev (1998) use this measure to evaluate traditional GARCH
specifications.
Although the research on multivariate GARCH models has produced a wide variety of models and
specifications, these have not yet been successful in financial applications as they have not been
capable of generalization to large covariance matrices. As computation becomes cheaper, and new
parsimonious models are formulated, the potential for building ever larger time varying conditional
covariance and correlation matrices increases. Models such as the vec and BEKK model of Engle
and Kroner (1995) have attractive properties as linear systems. The constant conditional correlation
(CCC) model of Bollerslev (1990) has the attraction of computational simplicity. A new model
called Dynamic Conditional Correlation (DCC) by Engle (2002) combines some of these features
to introduce a parsimonious correlation model to go with a conventional volatility model. Engle
and Sheppard (2001) estimate and test models of up to 100 assets. Ledoit and Santa-Clara (1998)
combine bivariate models to form multivariate models in a way which can be greatly expanded.
Correlation models can be estimated directly on intraday data. However as the frequency
increases, the asynchronicity of trades and returns leads to a serious underestimate of comovements.
This has been observed since Epps (1979) and the solutions of Scholes and Williams (1977) are
widely employed in spite of both theoretical and empirical difficulties. These are not appropriate
for ultra high frequency data and new solutions must be found.
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
428 R. ENGLE
range of the disturbance will be different for every observation. The variance and other higher
moments are unlikely to be constant. Efficient estimation via Maximum Likelihood is going to
be very difficult, although least squares will remain consistent. The probability of a near zero is
given by
Pt1 xt < D Pt1 εt < t
hence the error distribution must be discontinuous at t in order to satisfy (1).
The second conventional solution is to take logs. The model might then be written as
logxt D mt C ut 4
where
t D emt Eeut , t2 D e2mt Veut 5
This solution will not work if there are any exact zeros in fxt g. Sometimes a small constant is
added to eliminate the zeroes. However, this is more of a theoretical solution than a practical
one since the finite sample estimates are typically heavily influenced by the size of this constant.
Furthermore, the assumption in (1) that observations very near zero are possible, requires that
Pu < A > 0, for all A > 0. This is only true of very peculiar distributions.
Estimation of (4) requires the specification of both m and u. Clearly the relation between
mt and t depends upon the distribution of u. Thus even one-step forecasts require knowing the
distribution of u.
it has non-negative support. An exponential random variable with mean one is called a unit
exponential. Assuming that the disturbance is a unit exponential, then the univariate log likelihood
is simply
T
xt
L D logt 8
tD1
t
where theta represents the vector of parameters to be estimated. The first-order conditions for a
maximum of this likelihood function are
∂L
T
xt t ∂t
D 9
∂ tD1
2t ∂
By the law of iterated expectations, the expected value of the first-order condition when evaluated
at the true parameter value will be zero regardless of whether the density of x is truly unit
exponential. This implies that the log likelihood in (8) can be interpreted as a Quasi Likelihood
function and that parameters that maximize this are QMLE. Application of the theorem originally
given in White (1980) requires regularity conditions on the mean function and its determinants,
and gives general expressions for the covariance matrix.
A fairly general class of mean functions can be entertained for this problem. Suppose the mean
is linear in lagged x and in a k ð 1 vector of predetermined or weakly exogenous variables zt .
Then a (p, q) mean specification would be
p
q
0
t D ω C ˛j xtj C ˇj tj C zt 10
jD1 jD1
The parameters of this model may be restricted to ensure positive means for all possible
realizations, and to ensure stationary distributions for x. If z are positive variables, then sufficient
conditions for non-negativity are clearly that all parameters are positive. However, these are not
necessary. See Nelson and Cao (1992) for an examination of sufficient conditions. Sufficient
conditions for the covariance stationarity of x from Bollerslev, Engle and Nelson (1994) are that
z is covariance stationary and
p
q
˛j C ˇj < 1 11
jD1 jD1
This result can be formalized from Engle and Russell (1998) based upon a theorem in Lee and
Hansen (1994) for GARCH models. In their case, x is the duration between successive events, but
the theorem applies to any non-negative process. In this theorem, the process is assumed to be a
first-order GARCH type model possibly with unit or explosive roots.
If
Then:
(a) The maximizer of L will be consistent and asymptotically normal with a covariance matrix
given by the familiar robust standard errors as in Lee and Hansen.
p
(b) The model can be estimated with GARCH software by taking x as the dependent variable
and setting the mean to zero.
(c) The robust standard errors of Bollerslev and Wooldridge (1992) coincide with those in Lee
and Hansen.
From this corollary it is apparent that even mildly explosive models may be estimated consistently
by QMLE. From an examination of the mean specification in (10) it is apparent that the (p,q)
version of this MEM model with exogenous variables can also be estimated using GARCH software
p
by making xi the dependent variable, specifying it to have zero mean and an error process
assumed normal GARCH(p, q) with exogenous variables z. The estimated ‘conditional variance’
is then the conditional mean of x. Multi-step forecasts of x are computed simply by multi-step
forecasts of the conditional variance.
The clear advantage of the exponential error assumption is that estimation is consistent regardless
of the correctness of this distribution. The disadvantage is that it is not fully efficient. However,
it is perfectly straightforward to formulate more general likelihood functions which allow more
flexible shapes of density function or time variation in higher moments of this density function.
rt D ht εt
rt2 D ht ε2t 12
In the squared version, the dependent variable is non-negative with mean h and a non-negative
multiplicative i.i.d. error with unit mean. This can be estimated directly by taking the absolute
value of returns as the dependent variable of a GARCH model.
1 Although the dependent variable does not need to be signed, the lagged variables in the conditional variance can still
include sign information if asymmetric models are sought.
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
432 R. ENGLE
The second paper in this direction is the ACD model of Engle and Russell (1998) where the
dependent variable is modelled as the time between events. The model proposed is
xi D i εi
p
q
0
i DωC ˛j xij C ˇj ij C zi 13
jD1 jD1
essentially a GARCH(p, q) with exogenous variables for the square root of durations.
Manganelli (2000) has used multiplicative error models for volume in a transactions model of
market microstructure. He estimated models of returns, duration and volume as a trivariate system
of equations and then examined impulse responses through this system.
Engle and Gallo (2002) and Chou (2001) estimated models on realized volatility and high low
ranges to obtain new more efficient volatility estimators. These models all had this form.
I will now present some illustrative results using realized volatilities of dollar/DM exchange
rates from Andersen et al. (2001). They construct a series of daily variances from squaring and
averaging 5-minute returns obtained from Olsen and associates for a period from 12/1986 to
4/1996. For almost 10 years of daily data, we have a return and a ‘realized variance’ and we want
to use these to model volatilities.
The data in Table I show that the average of the squared returns is close to the average realized
variance. The squared return, however, has a much larger range and a larger standard deviation.
Estimation of a GARCH(1,1) model with robust standard errors gives the rather conventional
results in Table II.
Introducing lagged realized variance, v(1) into this equation produces the results in Table III.
Realized variance does have explanatory power beyond past squared returns to predict squared
returns. In fact, lagged squared returns are no longer significant in this model.
Now I will estimate a model for realized volatility which is the square root of realized variance
(Table IV). Again, a GARCH(1,1) will be specified but in this case it should be recognized that
this is a multiplicative error model with an exponential error assumption. Standard errors are
R2 V
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
NEW FRONTIERS FOR ARCH MODELS 433
computed using the Bollerslev–Wooldridge (1992) formulation. Two models are estimated; the
second also includes past squared returns.
As can be seen, the first equation is similar to a conventional GARCH model except that the
coefficient on the ARCH term is much larger and the coefficient on the GARCH term much smaller
than usual. This process is still highly persistent but is more volatile; it has a higher volatility of
volatility. This will be discussed in much detail later in this paper. In the second set of results, the
previous squared return is introduced. It has a small coefficient but a very small standard error2
and is quite significant in relation to the asymptotic normal distribution. In fact, the coefficient is
quite similar to that in Table II. There is apparently information in the squared return that helps
to predict realized volatility.
This estimator is potentially inefficient as it assumes an exponential density when another
density could be better. In fact the squared standardized residuals which in this case are
estimates of the disturbances, have a mean of one but a mode which is close to one as
well and a standard deviation which is 0.75, revealing underdispersion. A plot is given in
Figure 1. A more flexible set of density functions for non-negative disturbances is the gamma
density, which is a continuous version of a Chi Square. Setting the mean to unity leaves
one shape parameter in the density, which is the degrees of freedom/2 in a Chi Square.
The results are in Table V for both specifications. These densities are strongly preferred over
the exponential, which achieved a log likelihood of only 619. A chi square of 7 degrees
of freedom divided by 14, which is its mean, has a plot similar to that in Figure 1. The
significance of the squared returns is supported again with very similar coefficients and standard
errors.
2 Interestingly, when conventional standard errors are used, the t-statistic is only 1. It is not clear why the robust standard
errors are so much smaller.
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
434 R. ENGLE
1.0
0.8
0.6
0.4
0.2
0.0
0 1 2 3 4 5 6 7 8 9 10 11
RESID01^2
Figure 1. Estimated residual density function
Many further extensions are possible including more flexible densities and time varying
densities. Preliminary efforts to estimate time varying gamma densities for this data set indicated
that the assumption of i.i.d. gamma was quite acceptable. For other data, this may not be
the case.
The potential applications of this set of statistical methods are enormous. As experience grows
with various types of data it will become clearer which densities will be most useful and how best
to parameterize any changing density. The forecast performance of such models can be investigated
with both one-step and many-step forecasts (Table V).
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
NEW FRONTIERS FOR ARCH MODELS 435
find
EyT jz1,t , . . . , zk,t , where yT D urtC1 , rtC2 , . . . , rtCT
15
zj,t D vj rt , rt1 , . . . , rtM , j D 1, . . . , k
A potential solution to this problem is a simulation of simulations, which is, unfortunately,
unworkable. This consists of simulating the process from time 1 to t and computing the z’s. For
each of these simulations, a set of future r’s is simulated and the resulting yT ’s are averaged to
estimate the expectation given the z’s. Then the expected values of y given z are tabulated, or
approximated by a generous parameterization or a non-parametric regression. If N replications are
needed, this requires N2 simulations which is typically too time consuming.
The alternative strategy is Least Squares Monte Carlo as proposed by Longstaff and Schwartz
(2001). Their approach requires only N simulations and is thus computationally far more efficient.
They used the method to price American-style options by simulation. This has been thought to be
impossible because it requires knowing the value of the option before the expiration point—hence
the need for simulations of simulations. Their method can be used to find the early exercise frontier
for the options and then price these options.
The Least Squares Monte Carlo method estimates a cross-sectional regression with each
simulation being one observation with the dependent variable yT . The independent variables are
a non-parametric or generously parameterized function of the zj,t ’s. The result is a consistent
estimate of (15) by the standard properties of ordinary least squares. Since each observation is
independent, then the efficiency of this estimation will depend upon the distribution of the errors.
If in addition
VyT jz1,t , . . . , zk,t D s2 z1,t , . . . , zk,t 16
where s is a known function, then weighted least squares will be the best linear unbiased estimator
of this model.
This solution is closely related to the method of ‘reprojection’ introduced by Gallant and Tauchen
(1998) to find properties of simulated series. Their approach, however, initially constructs a very
long time series and thus does not deal easily with overlapping observations.
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
436 R. ENGLE
VoV D Vlog vT/t log vT/t1 ³ V[Et logvT Et1 logvT ] 18
This measure depends on the conditional variances but is calculated unconditionally over all
realizations. A conditional version could also be examined.
A particular application of this formula is to implied volatilities, which, under some conditions,
can be interpreted as the square root of a forecast of variance from today to the expiration of the
option at T. This measure can be constructed for any time series on implied volatilities; there may
be minor issues associated with the rollover from one contract to another, as maturity will not be
exactly constant. Such a number can be compared with forecasts from GARCH or other volatility
models. However, it is not easy to calculate this value for most volatility models. For example,
the VIX implied volatility series is approximately the 22-day Black–Scholes at-the-money implied
volatility of the S&P100 options. The estimated VoV for the sample period of the 1990s, as used
below, is
VoVVIX D stdev logVIX/VIXt 1 D 0.060
The standard error of an autoregression in log(VIX ) is almost the same number.
To apply the least squares Monte Carlo method to estimation of the VoV for a process, estimate
a cross-sectional regression:
yT,i D frt,i , zt1,i C εi 19
where 1/2
k
yT,i D log 2
rn,tCj 20
jD1
3 This approximation is exact if the Jensen inequality terms cancel. This will be true in a second-order Taylor expansion
if the coefficient of variation is the same at t and t 1.
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
NEW FRONTIERS FOR ARCH MODELS 437
is the log realized standard deviation of returns in the ith simulation. This is the log volatility
regression. An alternative is the variance regression when logs are not taken. The regressors are
returns at an intermediate time t and a set of variables based on information prior to time t. The
fitted value is an estimate of the expectation of y made at time t. A similar regression can be
estimated with information only to time t 1. Let this be
yT,i D gzt1,i C εi 21
The difference between f and g is the improvement in forecast attributed to information on rt . It
is the vector of innovations and its standard deviation is the estimated VoV.
VoV D fO
O g 22
The plot of f against r is interpreted as the News Impact Curve of Engle and Ng (1993) for
the statistic yT . Engle and Ng considered only the one-step-ahead conditional variance but this
can be extended to many statistics. Clearly the NIC will depend in general on the values of z,
and therefore it is common to evaluate this plot at an average value of z although any values can
be used.
The VoV in (22) is most easily calculated as the standard deviation of the difference between
the residuals of (19) and (21). To form an asymptotic confidence interval, let u be the squared
difference between the two residuals. Since each residual has mean zero and observations are
independent, applying the central limit theorem and the delta method,
N1/2 u1/2 VoV ! N0, u2 /VoV2 23
N!1
The design parameters are given in Table VI. There are some models with unusually high alphas
and others with very low alpha but high persistence as measured by the sum of alpha and beta.
Horizons of one month (22 days) and one year (252 days) are considered with 100,000
simulations. The log volatility equation is estimated for each of the 11 models to compute the
volatility of volatility. The volatility of volatility is given in Table VI along with the standard
errors of the estimates from (23).
It is interesting to note that at the monthly horizon, the values of volatility of volatility are
largest for the models with larger ˛. However, at the annual horizon, the most persistent models
generally are the most volatile. The standard errors in all cases are quite small. The daily VoV’s
of the monthly forecasts from the models we are accustomed to seeing, are well below the 0.06
for the option data.
It is easy to see that if the exponent on the innovation term is 1, it is simply a GARCH(1,1). The
estimated parameters are in Table VII.
4 The log likelihood for these three models is: 3741.492, 3730.631, and 3737.322 respectively. The symmetric
p
EGARCH is the same as the conventional EGARCH but omits the term rt / ht .
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
NEW FRONTIERS FOR ARCH MODELS 439
A second model is a non-linear GARCH model that is a hybrid between an integrated model
and a mean reverting model. This integrated model has no intercept and therefore will implode
eventually as shown by Nelson (1990). However, if the variance gets too small, it converts to
a mean reverting GARCH in a piecewise linear fashion. Thus it behaves like an integrated
GARCH (or Riskmetrics style Exponentially weighted moving average model) for large conditional
variances and like a conventional GARCH for small series. The formula is:
where I is an indicator function. Notice that it is an IGARCH for variances larger than delta, and
mean reverting below delta. The estimated values are given in Table VIII.
The daily variance of the S&P is about 1%. The estimated value of delta is not very significant
but is considerably above the mean variance indicating that the model is primarily mean reverting
except for large volatilities.
Several existing models were also considered such as the power model of Engle and Bollerslev
(1986):
p
htC1 D ω C ˛rt C ˇht 28
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
440 R. ENGLE
made on the innovations. This model might appear explosive since alpha and beta sum to more
than unity, however that configuration implies explosive behaviour only when p D 2.
Ding, Granger and Engle (1993) proposed the Power ARCH or PARCH model. In its symmetric
form it is given by
2
htC1 D ω C ˛rt C ˇht 29
and the estimated parameters are in Table X. In this case the exponent applies to all the variables
and is estimated at close to 1/2, suggesting a standard deviation style model.
A model written entirely in square root form was proposed by Taylor (1986) and used by
Schwert (1989). It is defined by
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
NEW FRONTIERS FOR ARCH MODELS 441
This wide range of GARCH type models have monthly VoV that range from 0.016 to 0.033 per
day. These would correspond to annualized volatilities from 25% to 50%, although one might not
want to annualize these in the conventional way since volatilities are mean reverting. The most
volatile models are the Taylor/Schwert, the PARCH and the EGARCH. The annual volatilities are
considerably lower ranging from 0.002 to 0.02. The largest VoV are the NLGARCH, the PARCH
and Taylor/Schwert.
Table XIII calculates the same measures when the simulated data are generated with non-
Gaussian innovations. Shocks are generated as t(6), t(4) and crash, where a crash innovation
is standard Gaussian with probability 0.995 and Gaussian (5,2) otherwise. The fatter-tailed
innovations increase the VoV but not dramatically. If the implied volatility series VIX is a standard
for the monthly volatility forecast, then all of these models are too smooth. The VoV of the VIX
is 0.06, which exceeds all of these cases. Of course, the VIX refers to the S&P100 which is more
volatile than the S&P500, and it has variability attributable to time variation in risk premia and
mispricing. Either the multi-step forecasts of these models are insufficiently responsive to news,
or options prices incorporate important additional sources of volatility.
The techniques developed for the conditional volatility models can be extended to any volatility
model that can be simulated. In this section, we develop this approach and apply it to several
types of stochastic volatility model. Several recent surveys of this material should be examined
by the interested reader: The main difference, is that the past information set can no longer be
summarized by a conditional variance, since this is no longer defined by the model. Instead, a
more general conditioning data set is introduced.
The natural generalization of equation (24) considers simply long lag distributions of absolute
and squared returns. Thus the equations that are estimated are:
50
logO T D c C ˛1 jrt j C ˛2 rt2 C ˛3 jrt j C 2
ˇ1j jrtj j C ˇ2j rtj 31
jD1
third is a long-memory stochastic volatility model. The generating equation for the SV model is
rt D t εt
logt D C logt1 C t
εt , vt ¾ IN0, I 32
as originally proposed by Taylor (1982) and later Nelson (1988). When this model is estimated
by Bayesian methods using the approach of Kim, Shephard and Chib (1998), the parameters are
given by the middle column of Table XIV.
When the same model was estimated using Shephard’s Maximum Likelihood routine, almost
identical parameters were obtained. For a different sample period of the S&P, Jaquier, Poulson
and Rossi (1994) (JPR) obtained similar parameter values given in the last column (Sandmann
and Koopman, 1998; Shephard and Pitt, 1997).
The ‘breaking volatility’ model is also a type of stochastic volatility model since it models the
volatility as a latent variable. In this model, volatility is constant until a random event occurs. At
this time a new volatility is drawn from a fixed distribution. The process can be defined as:
rt D t εt
2
2 t1 with probability p
t D
expt C otherwise
εt , t ¾ IN0, I 33
Two sets of parameters were used for this process based on a very casual matching of moments
(Table XV).
The third process is the Long Memory Stochastic Volatility model as proposed by Breidt, Crato
and de Lima (1998) and studied by Deo and Hurvich (2000). It is an alternative to the FIGARCH
Model of Baillie and Mikkelson (1996).
p 0.99 0.999
Kappa 1 1
Mu 0.5 0.5
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
NEW FRONTIERS FOR ARCH MODELS 443
Parameter Value
D 0.47
Kappa 0.6
Beta 0.4
p
rt D ˇ exp ut εt
1 Bd ut D t
εt , t ¾ IN0, I 34
Parameter estimates used for this model are shown in Table XVI.
The results from estimating the VoV for these five models are given in Table XVII.
From these results it appears that all these models have moderately high volatility of volatility
for long forecasts. Presumably this is because the persistence is governed by a different parameter
from the volatility of volatility. A related point has been made by Carnero, Pena and Ruiz (2001) in
the context of the stochastic volatility model. The difficulty of estimation and forecasting for these
models has made it impossible to observe this property before now. The long memory models
have the property that the long maturity VoV is high and rather closer to that of the shorter
horizon. The breaking volatility models are not systematically calibrated to the data and therefore
it is not clear which is to be considered most representative of this class of models. However,
the more frequently breaking model does have the highest VoV of all the Gaussian models at the
monthly horizon.
The use of the least squares Monte Carlo method can illuminate these and many other features of
time series models. In fact the applications of this general structure seem unlimited. The forecast
equation for these stochastic volatility models can be estimated, the expected payoff or utility
from path dependent hedging strategies can be assessed, the structure of derivative prices can
be analysed, and many other tasks can be rather simply evaluated, reducing the need to have
analytically simple models in order to obtain closed form forecasting results.
6. CONCLUSIONS
ARCH models have come a long way from 1982 to the present. In some respects their very success
has made it less interesting to continue research on volatility models. There is a sense in which this
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
444 R. ENGLE
is a known body of econometrics. In other ways, new horizons are opening all the time. The five
frontiers discussed in this paper are all fresh new areas of research with important applications and
lots of room for both theoretical and empirical developments. These five areas—high frequency
volatility, high dimension correlation, derivative pricing, modeling non-negative processes, and
analysing conditional simulations by Least Squares Monte Carlo—could well occupy our research
efforts for another decade, and would pay handsome dividends in terms of useful and reliable
research and practitioner tools. But probably, even brighter frontiers lie ahead and we can only
await them with anticipation.
REFERENCES
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
NEW FRONTIERS FOR ARCH MODELS 445
Ding Z, Granger CWJ, Engle RF. 1993. A long memory property of stock market returns and a new model.
Journal of Empirical Finance 1: 83–106.
Duan JC. 1995. The GARCH option pricing model. Mathematical Finance 5: 13–32.
Duan JC. 1997. Augmented GARCH(p, q) process and its diffusion limit. Journal of Econometrics 79(1):
97–127.
Engel C, Frankel JA, Froot KA, Rodrigues AP. 1995. Tests of conditional mean-variance efficiency of the
U.S. stock market. Journal of Empirical Finance 2: March, 3–18.
Engle RF. 1982. Autoregressive conditional heteroscedasticity with estimates of the variance of United
Kingdom Inflation. Econometrica 50(4): 987–1007.
Engle RF. 2000. The econometrics of ultra high frequency data. Econometrica 1–22.
Engle RF. 2001. GARCH 101: An introduction to the use of ARCH/GARCH models in applied Econometrics,
(forthcoming). Journal of Economic Perspectives.
Engle RF, Ishida I. 2001. Modeling variance of variance: the square-root, the afine and the CEV GARCH
models. Unpublished manuscript.
Engle RF. 2002. Dynamic conditional correlation: a new simple class of multivariate GARCH models JBES
(forthcoming).
Engle RF, Bollerslev T. 1986. Modeling the persistence of conditional variances. Econometric Reviews 5:
1–50.
Engle RF, Kane A, Noh J. 1994. Forecasting volatility and option prices of the S&P 500 Index. Journal of
Derivatives 217–30.
Engle RF, Lilien D, Robins R. 1987. Estimating time varying risk premia in the term structure: The ARCH-M
model. Econometrica 55: 391–407.
Engle RF, Kroner K. 1995. Multivariate simultaneous GARCH. Econometric Theory 11: 122–150.
Engle RF, Mustafa C. 1992. Implied ARCH models from options prices. Journal of Econometrics 52:
289–311.
Engle RF, Ng V. 1993. Measuring and testing the impact of news on volatility. Journal of Finance 48:
1749–1778.
Engle RF, Patton A. 2001. What good is a volatility model? Quantitative Finance 237–245.
Engle RF, Rosenberg J. 2000. Testing the volatility term structure using option hedging criteria. Journal of
Derivatives 8: 10–28.
Engle RF, Russell JR. 1998. Autoregressive conditional duration: a new model for irregularly spaced
transaction data. Econometrica 66: 1127–1162.
Engle RF, Shephard K. 2001. Theoretical and empirical properties of dynamic conditional correlation
multivariate GARCH. UCSD Discussion Paper 2001-15, NYU Discussion Paper.
Engle RF, Gallo E. 2002. A model for intra-daily volatility with multiple indicators. Unpublished manuscript.
Epps TW. 1979. Comovements in stock prices in the very short run. Journal of the American Statistical
Association 74(366): 291–298.
French KR, Schwert GW, Stambaugh RF. 1987. Expected stock returns and volatility. Journal of Financial
Economics 19: 3–30.
Gallant AR, Tauchen GE. 1998. Reprojecting partially observed systems with application to interest rate
diffusions. Journal of the American Statistical Association 93: 10–24.
Glosten LR, Jaganathan R, Runkle D. 1993. On the relation between the expected value and the volatility
of the nominal excess return on stocks. Journal of Finance 48: 1779–1801.
He C, Terasvirta T. 1999a. Fourth moment structure of the GARCH (p.q) process. Econometric Theory 15:
824–846.
He C, Terasvirta T. 1999b. Properties of moments of a family of GARCH processes. Journal of Econometrics
92: 173–192.
Heston S. 1993. A closed form solution for options with stochastic volatility with applications bond and
currency options. Review of Financial Studies 6: 327–343.
Higgins ML, Bera AK. 1992. A class of nonlinear ARCH models. International Economic Review 33: No. 1,
137–158.
Jacquier E, Polson NG, Rossi PE. 1994. Bayesian analysis of stochastic volatility models. Journal of Business
and Economic Statistics 12: 371–417.
Kim Sangjoon, Shephard N, Chib S. 1998. Stochastic volatility: likelihood inference and comparison with
ARCH models. Review of Economic Studies 65: 361–393.
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)
446 R. ENGLE
Ledoit O, Santa-Clara P. 1998. Estimating large conditional covariance matrices with an application to
international stock markets. Unpublished manuscript.
Lee S, Hansen B. 1994. Asymptotic theory for the GARCH(1,1) quasi-maximum likelihood estimator.
Econometric Theory 10: 29–52.
Li WK, Ling S, McAleer M. 2002. Recent theoretical results for time series models with GARCH errors.
Journal of Economic Surveys 16: 245–269.
Ling S, McAleer M. 2002a. Necessary and sufficient moment conditions for the GARCH(r,s) and asymmetric
power GARCH(r,s) Models. Econometric Theory 18: 722–729.
Ling S, McAleer M. 2002b. Stationarity and the existence of moments of a family of GARCH processes.
Journal of Econometrics 106: 109–117.
Longstaff F, Schwartz E. 2001. Valuing American options by simulation: a simple least squares approach.
The Review of Financial Studies 14: No. 1 Spring, 113–147.
Lumsdaine R. 1996. Consistency and asymptotic normality of the quasi-maximum likelihood estimator in
IGARCH(1,1) and covariance stationary GARCH(1,1) models. Econometrica 64: 575–596.
Manganelli S. 2000. Volume, Volatility and the Price Impact of Trades. PhD Dissertation, UCSD.
McCurdy T, Stengos T. 1992. A comparison of risk premium forecasts implied by parametric and nonpara-
metric conditional mean estimators. Journal of Econometrics 52: 225–244.
Nelson DB. 1990. Stationarity and persistence in the GARCH(1,1) model. Econometric Theory 6: 318–334.
Nelson DB. 1991. Conditional heteroscedasticity in asset returns: a new approach. Econometrica 59(2):
347–370.
Nelson DB. 1998. Time Series Behavior of Stock Market Volatility and Returns. PhD thesis, MIT.
Nelson DB, Cao CQ. 1992. Inequality constraints in the univariate GARCH model. Journal of Business and
Economic Statistics 10(2): 229–235.
Poon S-H, Granger CWJ. 2002. Forecasting Volatility in Financial Markets: A Review, Strathclyde Uni-
veristy, Glasgow.
Rabemananjara R, Zakoian JM. 1993. Threshold ARCH models and asymmetries in volatility. Journal of
Applied Econometrics 8: No. 1, 31–49.
Ritchken P, Trevor R. 1999. Pricing options under generalized GARCH and stochastic volatility processes.
Journal of Finance 54(1): 377–402.
Rosenberg J, Engle RF. 2001. Empirical pricing kernels. Journal of Financial Economics (forthcoming).
Sandmann G, Koopman SJ. 1998. Estimation of stochastic volatility models via Monte Carlo maximum
likelihood. Journal of Econometrics 87: 271–301.
Scholes M, Williams JT. 1977. Estimating betas from nonsynchronous data. Journal of Financial Economics
5(3): 309–327.
Schwert GW. 1989. Why does stock market volatility change over time? Journal of Finance 44: 1115–53.
Shephard N, Pitt MK. 1997. Likelihood analysis of non-Gaussian measurement time series. Biometrika
Vol 84: 653–67.
Styan GPH. 1973. Hadamard products and multivariate statistical analysis. Linear Algebra and Its Applica-
tions 6: 217–240.
Taylor SJ. 1982. Financial returns modelled by the product of two stochastic processes- a study of daily sugar
prices, 1961–1979. In Time Series Analysis: Theory and Practice, Anderson OE (ed.). North-Holland:
Amsterdam; 203–226.
Taylor SJ. 1986. Modelling Financial Time Series. John Wiley: Chichester.
White H. 1980. A heteroskedasticity-consistent covariance matrix estimator and a direct test for heteroskedas-
ticity. Econometrica 48(4): May, 817–838.
Copyright 2002 John Wiley & Sons, Ltd. J. Appl. Econ. 17: 425–446 (2002)