Professional Documents
Culture Documents
Outline
Motivation
Thinning operation
INteger-valued AutoRegressive (INAR) processes
◮ INAR(1) processes
◮ INAR(p) processes
Motivation
Discrete time non-negative integer-valued time series → counting series
Motivation
Discrete time non-negative integer-valued time series → counting series
Examples:
the daily number of seizures of epileptic patients
the yearly number of plants in a region
the daily number of guest nights in a hotel
the monthly incidence of a disease.
Motivation
Discrete time non-negative integer-valued time series → counting series
Examples:
the daily number of seizures of epileptic patients
the yearly number of plants in a region
the daily number of guest nights in a hotel
the monthly incidence of a disease.
Usual linear time series models (ARMA processes) are not suitable.
Motivation
Discrete time non-negative integer-valued time series → counting series
Examples:
the daily number of seizures of epileptic patients
the yearly number of plants in a region
the daily number of guest nights in a hotel
the monthly incidence of a disease.
Usual linear time series models (ARMA processes) are not suitable.
Thinning operation
m
Multiplication counterpart on the integer-valued context.
Thinning operation
Binomial thinning operation [Steutel and Van Harn (1979)]
X : non-negative integer-valued random variable (r.v.), α ≥ 0
X
α ◦ X = ∑ Yj
j=1
{Yj } ⊆ N0 (counting series): sequence of independent and identically distributed
(i.i.d.) r.v., independent of X, P(Yj = 1) = 1 − P(Yj = 0) = α
⇓
α ◦ X is the number of successes, with probability α , in X trials
Thinning operation
Binomial thinning operation [Steutel and Van Harn (1979)]
X : non-negative integer-valued random variable (r.v.), α ≥ 0
X
α ◦ X = ∑ Yj
j=1
{Yj } ⊆ N0 (counting series): sequence of independent and identically distributed
(i.i.d.) r.v., independent of X, P(Yj = 1) = 1 − P(Yj = 0) = α
⇓
α ◦ X is the number of successes, with probability α , in X trials
Generalized thinning operation [Gauthier and Latour (1994); Silva and Oliveira (2004, 2005)]
X
α ∗ X = ∑ Yj
j=1
{Yj }: sequence of i.i.d.r.v., independent of X, with some discrete distribution such that
E[Yj ] = α , Var[Yj ] = σ 2 , E[Yj 3 ] = γ , E[Yj 4 ] = κ
Thinning operation PDMA-UP - October 2009 4 / 30
Isabel Silva Magalhães Integer-valued time series
Xt = α ∗ Xt−1 + et , t = 1, . . . , N
Xt = α ∗ Xt−1 + et , t = 1, . . . , N
0<α ≤1
{et } ⊆ N0 : sequence of i.i.d. discrete r.v. (arrival process)
counting series, {Yj }, are independent, and independent of {et }, and such that
Xt = α ◦ Xt−1 + et
⇓
{et } Poisson distributed and {Yj } Bernoulli distributed
Xt = α ◦ Xt−1 + et
⇓
{et } Poisson distributed and {Yj } Bernoulli distributed
Xt = α ◦ Xt−1 + et
⇓
{et } Poisson distributed and {Yj } Bernoulli distributed
λ
X1 ∼ Po 1−α λ
=⇒ {Xt } ∼ Po
{et } ∼ Po(λ ) 1−α
Xt = α1 ∗ Xt−1 + · · · + αp ∗ Xt−p + et , t = 1, . . . , N,
αi ∗ Xt−i = ∑j=0
t−i X
Yj,i , i = 1, . . . , p,
Xt = α1 ∗ Xt−1 + · · · + αp ∗ Xt−p + et , t = 1, . . . , N,
αi ∗ Xt−i = ∑j=0
t−i X
Yj,i , i = 1, . . . , p,
INteger-valued Moving Average (INMA) model [Al-Osh and Alzaid (1988); Mckenzie (1988);
Brännäs and Hall (2001)]
Multivariate INAR(p) process [Franke and Subba Rao (1995); Latour (1997)]
Poisson INAR(1) process is a M/M/∞ queueing system observed at regularly spaced interval of
times [Steutel et al. (1983); Mckenzie (1988)]
INAR(p) is a Multitype Branching processes with immigration, BGWI(p) [Dion et al. (1995)]
Time domain
Method of Moments
◮ Second-order: Yule-Walker estimation (YW)
◮ Third-order: Estimation using the Cumulant Third-Order Recursion equation
(TOR),
Least Squares (LS) estimation
◮ Second-order: Conditional Least Squares (CLS)
◮ Third-order: LS estimation based on high-order statistics (LS_HOS),
Frequency domain
Whittle estimation (WHT),
Time domain
Method of Moments
◮ Second-order: Yule-Walker estimation (YW)
◮ Third-order: Estimation using the Cumulant Third-Order Recursion equation
(TOR),
Least Squares (LS) estimation
◮ Second-order: Conditional Least Squares (CLS)
◮ Third-order: LS estimation based on high-order statistics (LS_HOS),
Frequency domain
Whittle estimation (WHT),
YW, CLS, WHT, TOR, LS_HOS: do not assume the Poisson distribution for the
arrival process → more adaptive and flexible methods
Then
N 1/2 (α̂ − α ) is AN(0p , Vyw ),
Theorem (Cumulant Third-Order Recursion (TOR) equation of INAR(p) processes [Silva (2005)])
Let {Xt } be a stationary INAR(p) process. Then the third-order cumulants of Xt can
be written in a single cumulant Third-Order Recursion (TOR) equation by
Theorem (Cumulant Third-Order Recursion (TOR) equation of INAR(p) processes [Silva (2005)])
Let {Xt } be a stationary INAR(p) process. Then the third-order cumulants of Xt can
be written in a single cumulant Third-Order Recursion (TOR) equation by
~
w
CX (−k, −k) = CX (0, k)
CX (0, 0) CX (1, 1) ··· CX (p − 1, p − 1) α1 CX (0, 1)
CX (0, 1) CX (0, 0) ··· CX (p − 2, p − 2) α2 CX (0, 2)
C3,X α = . . .. . . = . = c3,X
. . . . . .
. . . . .
CX (0, p − 1) CX (0, p − 2) ··· CX (0, 0) αp CX (0, p)
1 M−k (i)
M ∑j=1 j
(i) (i) (i) (i)
ĈX (k, k) = (X − X )(Xj+k − X )2 , k = 0, . . . , p − 1,
1 M−k (i)
ĈX (0, k) = ∑j=1 (Xj − X )2 (Xj+k − X ),
(i) (i) (i) (i)
k = 1, . . . , p,
M
1 B (i)
B ∑i=1 X
ĈX (k, k) = Ĉ (k, k), k = 0, . . . , p − 1,
1 B (i)
ĈX (0, k) = ∑i=1 ĈX (0, k), k = 1, . . . , p,
B
α = ĉ3,X .
Solve Ĉ3,X α̂
1 M−k (i)
M ∑j=1 j
(i) (i) (i) (i)
ĈX (k, k) = (X − X )(Xj+k − X )2 , k = 0, . . . , p − 1,
1 M−k (i)
ĈX (0, k) = ∑j=1 (Xj − X )2 (Xj+k − X ),
(i) (i) (i) (i)
k = 1, . . . , p,
M
1 B (i)
B ∑i=1 X
ĈX (k, k) = Ĉ (k, k), k = 0, . . . , p − 1,
1 B (i)
ĈX (0, k) = ∑i=1 ĈX (0, k), k = 1, . . . , p,
B
α = ĉ3,X .
Solve Ĉ3,X α̂
Theorem (Asymptotic distribution of the CLS estimators of an INAR(p) process [Latour (1998)])
Let {X1 , . . . , XN } be observations of an INAR(p) process such that E [e3t ] < ∞ and
(k)
E [(Yi )3 ] < ∞, for k = 1, . . . , p and k ∈ N. Let θ = [ α1 · · · αp µe ]T and let θ̂
denote the Conditional Least Squares estimator of θ , given by
( )
N
QN (θ̂ ) = min{QN (θ )} = min
θ θ
∑ (Xt − α1 Xt−1 − . . . − αp Xt−p − µe ) 2
.
t=p+1
Then
√
N(θ̂ − θ ) ∼ N (0p+1 , V−1 WV−1 ),
where 0p+1 is a (p + 1) × 1 vector of zeros and the entries of V and W are defined by
µX (i − j) i, j = 1, . . . , p, p
i = p + 1 or j = p + 1, i 6= j, [W]i,j = σe [V]i,j + ∑ σk µX (k − i, k − j).
2 2
[V]i,j = µX
k=1
1 i = j = p + 1,
m
µX (0, 1) µX (0, 0) µX (1, 1) ... µX (p − 1, p − 1) α1 1
µX (0, 2) µX (0, 1) µX (0, 0) ... µX (p − 2, p − 2) α2 1
. = . . .. . . + µe µX (0) .
. . . . . . .
. . . . . .
µX (0, p) µX (0, p − 1) µX (0, p − 2) ... µX (0, 0) αp 1
p p
µX (0) = ∑ αi µX (i) + µe µX + Vp , with Vp = σe 2 + µX ∑ σi 2
i=1 i=1
µ 3,X = Hθ
µ 3,X = Hθ
θ may be estimated by least squares, ie, minimizing the squared error between µ 3,X
and the third-order moments of the data:
µ 3 = [ µ (0, 1) · · · µ (0, p) ]T
µ 3,X = Hθ
θ may be estimated by least squares, ie, minimizing the squared error between µ 3,X
and the third-order moments of the data:
µ 3 = [ µ (0, 1) · · · µ (0, p) ]T
µ 3,X = Hθ
θ may be estimated by least squares, ie, minimizing the squared error between µ 3,X
and the third-order moments of the data:
µ 3 = [ µ (0, 1) · · · µ (0, p) ]T
2
1 N
−iωj t f (ωj ) 2
IN (ωj ) = ∑ Xt e ∼ χ2 ,
2π N t=1 2
2
1 N
−iωj t f (ωj ) 2
IN (ωj ) = ∑ Xt e ∼ χ2 ,
2π N t=1 2
2
1 N
−iωj t f (ωj ) 2
IN (ωj ) = ∑ Xt e ∼ χ2 ,
2π N t=1 2
µ̂e = X 1 − ∑i=1 α̂i , σ̂e2 = V̂p − X ∑i=1 σ̂i2 ,
p p
where V̂p = R̂(0) − ∑pi=1 α̂i R̂(i) and σ̂i2 is an estimator of Var[Yj,i ], i = 1, . . . , p. For
instance, σ̂i2 = α̂i (1 − α̂i ), for the binomial thinning operation.
µ̂e = X 1 − ∑i=1 α̂i , σ̂e2 = V̂p − X ∑i=1 σ̂i2 ,
p p
where V̂p = R̂(0) − ∑pi=1 α̂i R̂(i) and σ̂i2 is an estimator of Var[Yj,i ], i = 1, . . . , p. For
instance, σ̂i2 = α̂i (1 − α̂i ), for the binomial thinning operation.
Simulation results
Non-admissible estimates → constrained estimation (CLS, WHT, LS_HOS)
µ̂e = X 1 − ∑i=1 α̂i , σ̂e2 = V̂p − X ∑i=1 σ̂i2 ,
p p
where V̂p = R̂(0) − ∑pi=1 α̂i R̂(i) and σ̂i2 is an estimator of Var[Yj,i ], i = 1, . . . , p. For
instance, σ̂i2 = α̂i (1 − α̂i ), for the binomial thinning operation.
Simulation results
Non-admissible estimates → constrained estimation (CLS, WHT, LS_HOS)
Sample bias, variance, mean square error and univariate skewness decrease as
the sample size increases ⇒ consistency and symmetry
µ̂e = X 1 − ∑i=1 α̂i , σ̂e2 = V̂p − X ∑i=1 σ̂i2 ,
p p
where V̂p = R̂(0) − ∑pi=1 α̂i R̂(i) and σ̂i2 is an estimator of Var[Yj,i ], i = 1, . . . , p. For
instance, σ̂i2 = α̂i (1 − α̂i ), for the binomial thinning operation.
Simulation results
Non-admissible estimates → constrained estimation (CLS, WHT, LS_HOS)
Sample bias, variance, mean square error and univariate skewness decrease as
the sample size increases ⇒ consistency and symmetry
LS_HOS, WHT and CLS provides good results
◮ in terms of smaller sample bias, variance and mean square error
monthly counts
8
[Zeger (1988): Parameter-driven
6
model], and sample
4
autocorrelation and partial 2
autocorrelation functions 0
1970 1972 1974 1976 1978 1980 1982 1984
year
1.2 0.3
1
0.2
0.8
0.1
0.6
ρ(k)
φ(k)
0.4
0
0.2
−0.1
0
−0.2 −0.2
0 5 10 15 20 25 30 0 5 10 15 20 25 30
k k
Table: Parameter estimates of the INAR(1) model fitted to the polio data.
Application to real data PDMA-UP - October 2009 23 / 30
Isabel Silva Magalhães Integer-valued time series
45
The number of Swedish
40
mechanical paper and
35
Number of plants
pulp mills, from 1921 to 30
Hellström (2001): 15
Explanatory variables] 10
5
1920 1930 1940 1950 1960 1970 1981
MSE between the observations and the fitted models based on TOR_1B, LS_HOS and
CLS estimates
Application to real data PDMA-UP - October 2009 25 / 30
Isabel Silva Magalhães Integer-valued time series
50
Real data
45 CLS
LS_HOS
40
Number of plants
30
the fitted values
25
considering the LS_HOS
20
and CLS estimates
15
10
5
1920 1930 1940 1950 1960 1970 1981
Recent developments
Theme proposal
Theme proposal
Gomes, D., Canto e Castro, L., 2009. Generalized integer-valued random coefficient for a first order
structure autoregressive (RCINAR) process. Journal of Statistical Planning and Inference, vol. 139
(12), pp. 4088–4097.
Zheng, H., Basawa, I. V., Datta, S., 2007. First-order random coefficient integer-valued
autoregressive processes. Journal of Statistical Planning and Inference, vol. 137 (1), pp. 212–229.
References I
Journal of Time Series Analysis, vol. 8, pp. 261–275. Applied Stochastic Models in Business and Industry, vol.
17, pp. 277–291.
Al-Osh, M.A. and Alzaid, A.A., 1988.
Integer-valued moving average (INMA) process. Dion, J-P. and Gauthier, G. and Latour, A., 1995.
Statistical Papers, vol. 29, pp. 281–300. Branching processes with immigration and integer-valued
time series.
Al-Osh, M.A. and Alzaid, A.A., 1991. Serdica Mathematical Journal, vol. 21, pp. 123–136.
Binomial autoregressive moving average models.
Communications in Statistics: Stochastic Models, vol. 7, Du, Jin-Guan and Li, Yuan, 1991.
Journal of Applied Probability, vol. 27, pp. 314–324. processes and their application to modelling epileptic
seizure counts.
Brännäs, K., 1994. In Developments in Time Series Analysis: in honour of
Estimation and testing in integer-valued AR(1) models. Maurice B. Priestley, Chapman & Hall, pp. 310–330.
Technical Report, Umeå University, Sweden, 335.
Franke, J. and Subba Rao, T., 1995.
Brännäs, K., 1995. Multivariate first order integer valued autoregressions.
Explanatory variables in the AR(1) count data model. Technical report, Math. Dep., UMIST.
Technical Report, Umeå University, Sweden, 381.
References II