Content deleted Content added
Marcocapelle (talk | contribs) removed Category:Time series analysis; added Category:Time series models using HotCat |
wtf was that |
||
(27 intermediate revisions by 18 users not shown) | |||
Line 1:
{{Short description|Method to find best fit of a time-series model}}
In [[time series analysis]], the '''Box–Jenkins method''',<ref>{{cite book |
==Modeling approach==
The original model uses an iterative three-stage modeling approach:
#''Model identification and [[model selection]]'': making sure that the variables are [[stationary process|stationary]], identifying [[seasonality]] in the dependent series (seasonally differencing it if necessary), and using plots of the [[autocorrelation|autocorrelation (ACF)]] and [[partial autocorrelation|partial autocorrelation (PACF)]] functions of the dependent time series to decide which (if any) autoregressive or moving average component should be used in the model.
#''[[Parameter estimation]]'' using computation algorithms to arrive at coefficients that best fit the selected ARIMA model. The most common methods use [[maximum likelihood estimation]] or [[non-linear least-squares estimation]].
#''[[Statistical model validation|
The data they used were from a gas furnace. These data are well known as the Box and Jenkins gas furnace data for benchmarking predictive models.
Commandeur & Koopman (2007, §10.4)<ref>{{cite book |
==
===Stationarity and seasonality===
The first step in developing a Box–Jenkins model is to determine
====Detecting stationarity====
Stationarity can be assessed from a [[run sequence plot]]. The run sequence plot should show constant ___location and [[Scale (ratio)|scale]]. It can also be detected from an [[autocorrelation plot]]. Specifically, non-stationarity is often indicated by an autocorrelation plot with very slow decay. One can also utilize a [[Dickey-Fuller test]] or [[Augmented Dickey-Fuller test]].
====Detecting seasonality====
Seasonality (or periodicity) can usually be assessed from an autocorrelation plot, a [[seasonal subseries plot]], or a [[spectral plot]].
====Differencing to achieve stationarity====
Box and Jenkins recommend the differencing approach to achieve stationarity. However, [[curve fitting|fitting a curve]] and subtracting the fitted values from the original data can also be used in the context of Box–Jenkins models.
Line 29 ⟶ 30:
===Identify ''p'' and ''q''===
Once stationarity and seasonality have been addressed, the next step is to identify the order (i.e. the ''p'' and ''q'') of the autoregressive and moving average terms. Different authors have different approaches for identifying ''p'' and ''q''. Brockwell and Davis (1991)<ref>{{cite book |
====Autocorrelation and partial autocorrelation plots====
Line 36 ⟶ 37:
Specifically, for an [[AR(1)]] process, the sample autocorrelation function should have an exponentially decreasing appearance. However, higher-order AR processes are often a mixture of exponentially decreasing and damped sinusoidal components.
For higher-order autoregressive processes, the sample autocorrelation needs to be supplemented with a partial autocorrelation plot. The partial autocorrelation of an AR(''p'') process becomes zero at lag ''p''
The autocorrelation function of a [[moving average model|MA(''q'')]] process becomes zero at lag ''q''
The sample partial autocorrelation function is generally not helpful for identifying the order of the moving average process.
Line 53 ⟶ 54:
| Autoregressive model. Use the partial autocorrelation plot to help identify the order.
|-
! One or more spikes, rest are essentially zero (or close to zero)
| [[Moving average model]], order identified by where plot becomes zero.
|-
Line 65 ⟶ 66:
| Include seasonal autoregressive term.
|-
! No decay to zero (or it decays extremely slowly)
| Series is not stationary.
|}
Hyndman & Athanasopoulos suggest the following:<ref>{{cite
:The data may follow an ARIMA(''p'',''d'',0) model if the ACF and PACF plots of the differenced data show the following patterns:
Line 82 ⟶ 83:
==Box–Jenkins model estimation==
Estimating the parameters for
==Box–Jenkins model diagnostics==
Line 93 ⟶ 94:
If these assumptions are not satisfied, one needs to fit a more appropriate model. That is, go back to the model identification step and try to develop a better model. Hopefully the analysis of the residuals can provide some clues as to a more appropriate model.
One way to assess if the residuals from the Box–Jenkins model follow the assumptions is to generate [[statistical graphics]] (including an autocorrelation plot) of the residuals. One could also look at the value of the [[
==References==
{{Reflist}}
* {{cite book |last=Pankratz |first=Alan |year=1983 |title=Forecasting with Univariate Box–Jenkins Models: Concepts and Cases |___location=New York |publisher=John Wiley & Sons }}▼
==Further reading==
* {{citation | title= Comparison of Box–Jenkins and objective methods for determining the order of a non-seasonal ARMA model | author1-first= S. | author1-last= Beveridge | author2-first= C. | author2-last= Oickle | journal= [[Journal of Forecasting]] | year= 1994 | volume= 13 | issue= 5 | pages= 419–434 | doi= 10.1002/for.3980130502}}
▲* {{
==External links==
* [https://web.archive.org/web/20070318000551/http://statistik.mathematik.uni-wuerzburg.de/timeseries/ A First Course on Time Series Analysis]
* [http://www.itl.nist.gov/div898/handbook/pmc/section4/pmc445.htm Box–Jenkins models] in the Engineering Statistics Handbook of [[NIST]]
* [http://robjhyndman.com/papers/BoxJenkins.pdf
* [http://support.sas.com/resources/papers/proceedings13/454-2013.pdf The
{{NIST-PD}}
{{Authority control}}
{{DEFAULTSORT:Box-Jenkins}}
|