In the statistical physics literature, heavy-tailed inter-arrival times are said to be the main signature of b burstyb dynamics. Such dynamics have been observed for financial time series, earthquakes, solar flares and neuron voltage spikes. This talk is concerned with the modelling of extremes of events when events follow such bursty dynamics. We assign i.i.d. magnitudes (marks) to the events in a heavy-tailed renewal process and apply the b Peaks Over Thresholdb method from Extreme Value Theory. Leveraging geometric stability, it can be shown that the threshold exceedance times asymptotically form a b fractional Poisson processb with Mittag-Leffler inter-arrival times. Finally, we discuss methods for inference on the tail and scale parameters of the bursty dynamics.
Vajna, S., TC3th, B., & KertC)sz, J. (2013). Modelling bursty time series. New Journal of Physics, 15(10), 103023. http://doi.org/10.1088/1367-2630/15/10/103023
Karsai, M., Kaski, K., BarabC!si, A. L., & KertC)sz, J. (2012). Universal features of correlated bursty behaviour. Scientific Reports, 2. http://doi.org/10.1038/srep00397
But: for renewal processes with \(1 < \beta < 2\) (finite mean, infinite variance) the ACF decays with exponent \(2-\beta\).
Quick fix: log-transform the waiting times.
a sequence of consecutive events where all inter-arrival times are \(< \Delta t\).
If the two plots are visually different, renewal assumption is false.
(Karsai et al 2012, Universal features of correlated bursty behaviour, Scientific Reports)
Clustering coefficients for a growing network, where nodes are preferentially attached to nodes with large clustering coefficients.
Bagrow, J. P., & Brockmann, D. (2013). Natural emergence of clusters and bursts in network evolution. Physical Review X, 3(2), 1b 6. http://doi.org/10.1103/PhysRevX.3.021016
| Waiting time | Magnitude |
|---|---|
| Server requests | filesize |
| Earthquakes | magnitudes |
| Email sent | email replied |
| Solar flares | magnitudes |
| Time between rest for paper currency | distance travelled |
| Time between raindrops | raindrop size |
| Time between price jumps | price jumps |
How does one model magnitude and timing of extreme events, if the underlying dynamics are a heavy-tailed renewal process?
Consider the uncoupled case: \(\{W_k\}\) and \(\{J_k\}\) are independent.
What is the distribution of
Renewal Process
\[N(t) = \{ n\ge 0: W_1 + \ldots + W_n \le t \}\]
counts events up to time \(t\).
CTRM Process:
\[M(t)=\bigvee_{k=1}^{N(t)}J_k=\max\{J_1, \ldots,J_{N(t)}\}\]
(s. dashed line)
First magnitude above a threshold \(\ell\) is:
\[\tau(\ell) = \min\{k: J_k > \ell\}\]
Exceedance Process:
\[T(\ell) = \sum_{k=1}^{\tau(\ell)} W_k\]
We have \(T(\ell) > t \Leftrightarrow M(t) \le \ell\).
(CTRM and Exceedance Time are inverse to each other.)
(s. dashed line)
\[p_\ell T(\ell) \stackrel{d}{\to} {\rm Exp}(\mu_W)\]
\[p_\ell^2 {\rm Var} T(\ell) \to \mu_W^2\]
where \(\mu_W = \mathbf E[W_k]\) (e.g.B Gut b 99).
Let \(a_n\) and \(b_n\) be such that \(F_J(a_n x + b_n)^n \to \Lambda(x)\) for all \(x\)
Let \(g(t) = t / [\Gamma(2-\alpha) \int_0^t (1-F_W(u))\,du]\).
Then the CTRM process \(M(t)\) satisfies
\[\lim_{t\to \infty} \mathbf P\left[M(t) \le a_{[g(t)]} x + b_{[g(t)]}\right] = E_\beta(\log \Lambda(x))\]
where \(\beta \in (0,1]\) is the tail index of \(W_k\) and \(E_\beta(\cdot)\) is the Mittag-Leffler function (more later).
As \(n \to \infty\), assume there are increasing norming functions \(a\), \(d\) and \(b\) such that
Then as \(c \to \infty\),
\[\frac{M(ct) - d(\tilde b(c))}{a(\tilde b(c))} \Rightarrow A(E(t))\]
Here
\[b(\tilde b(c)) \sim c \sim \tilde b(b(c))\]
(Meerschaert & Stoev b 08).
Other approaches:
These models have been studied thoroughlyb & but how do you fit them to data?
Define the Exceedance as
\[X(\ell) = M(T_\ell) - \ell.\]
In the uncoupled case, it follows easily that Exceedance \(X(\ell)\) and Exceedance Time \(T(\ell)\) are independent.
Asymptotically for large \(\ell\), \(X_\ell\) follows a generalized Pareto distribution:
\[X_\ell \sim GP(\xi, \bar \sigma)\]
Extract the red lines and fit GEV distributions to them!
For threshold \(\ell\) in \(\{J_{(1)} > J_{(2)} > \ldots > J_{(n)}\}\):
Then plot parameters estimates \(\hat \xi, \hat \sigma\) against \(k\).
Example: earthquake magnitudes shape parameterb &
b & and scale parameter:
By inspection, pick \(\hat \xi = -0.1\), \(\hat \sigma = 0.6\).
Small \(k\): few data (high variance) but asymptotically close to GP distribution (low bias)
High \(k\): many data (low variance) but asymptotically far away from GP distribution (high bias)
For threshold \(\ell\) in \(\{J_{(1)} > J_{(2)} > \ldots > J_{(n)}\}\):
Then plot the parameters of this distribution against \(k\).
But what is the distribution of \(T(\ell)\)??
\(T(\ell) =\) distances between blue circles.
\[E_\beta(z) = \sum_{k=0}^\infty \frac{z^k}{\Gamma(1+\beta k)}\]
where \(0 < \beta \le 1\).
\(E_\beta(z)\) nests the exponential function: \(E_1(z) = e^z\).
For \(\beta < 1\), \(E_\beta(-z^\beta)\) is \(\rm{RV}_\infty(-\beta)\).
A r.v. \(T_\beta\) is Mittag-Leffler-1 distributed with tail parameter \(\beta \in (0,1]\) if
\[\mathbf P(T > t) = E_\beta(-t^\beta), \quad t > 0.\]
A r.v. \(T_\beta\) is Mittag-Leffler-2 distributed with parameter \(\beta \in (0,1]\) if
\[\int_0^\infty e^{-st} \mathbf P(T \in dt) = E_\beta(-s).\]
We write \(\sigma T_\beta \sim {\rm ML2}(\beta, \sigma)\) for \(\sigma > 0\).
Let \(D(u)\) be a stable subordinator with Laplace Transform \(\exp(-us^\beta)\).
Let \(E(t) = \inf\{u: D(u) > t\}\) (inverse stable subordinator).
Then
\[E(t) \sim {\rm ML2}(\beta, t^\beta).\]
A random variable \(Y\) is geometric stable if and only if its characteristic function is
\[\psi(t) = \frac{1}{1 + \sigma^\alpha |t|^\alpha \omega_{\alpha, \beta}(t) - i\mu t}\]
where for \(\alpha \neq 1\)
\[\omega_{\alpha, \beta}(t)= \begin{cases} 1 - i\beta {\rm sign}(t) \tan(\pi \alpha / 2), & \alpha \neq 1 \\ 1 + i\beta(2/\pi) {\rm sign}(t) \log|t|, & \alpha = 1 \end{cases}\]
Write \(Y \sim {\rm GS}(\alpha, \beta, \sigma, \mu)\); (stability, skewness, scale, location).
(Parametrization by Samorodnitsky & Taqqu, see Kozubowski & Rachev, 1999)
A r.v. \(Y\) on \(\mathbb R\) is GS (geometric stable) if for \(N_p \sim {\rm Geom}(p)\) there exist i.i.d. r.v.b s \(X_1, X_2, \ldots\), independent of \(N_p\), a norming sequence \(A_p\) and centering constants \(b_p\) such that
\[A_p \sum_{k=1}^{N_p} (X_k + b_p) \stackrel{d}{\to} Y \text{ as } p \downarrow 0.\]
Then \(X_1\) is weakly geometrically attracted to \(Y\), and the collection of such \(X_k\) is called the generalized domain of geometric attraction (GDOGA) of \(Y\).
Let \(\beta \in (0,1]\). Then \(X \in \text{GDOGA}({\rm ML(\beta,1)})\) if and only if
\[1 - F_X(t) \in {\rm RV}_\infty(-\beta)\]
Recall:
Moreover, \(N_{p_\ell} := \min\{k: J_k > \ell\}\) is geometric, independent of \(W_k\), and
\[T(\ell) = \sum_{k=1}^{N_{p_\ell}} W_k.\]
Suppose the magnitudes \(J_k\) are supported on the interval \([x_L, x_R]\).
Suppose that \(\mathbf P(W_k > t) = L(t) t^{-\beta}\) for some slowly varying function \(L(t)\) and \(\beta \in (0,1)\).
Let \(\ell \in [x_L, x_R]\) be a threshold, and write \(p_\ell = 1 - F_J(\ell)\) for the probability that a magnitude exceeds threshold \(\ell\).
Then for some \(\sigma > 0\),
\[p_\ell^{1/\beta} T(\ell) \stackrel{d}{\to} {\rm ML}(\beta, \sigma), \quad \ell \uparrow x_R\]
Gnedenko (1972) predates the Shock-process literature (1988+); Yet geometric stability does not appear in the Shock process literature.
It took the end of the Cold War and the invention of the internet to put the two things together!
For large thresholds \(\ell\), we may approximate
\[T(\ell) \approx {\rm ML}(\beta, \sigma p_\ell^{-1/\beta}).\]
Small k: few data (high variance) but many geometric summands, hence asymptotically close to M-L distribution (low bias)
High k: many data (low variance) but few geometric summands, hence far away from GP distribution (high bias)
Let \(T \sim ML(\beta, \sigma)\).
(Cahoy et al b 10)
Recent algorithm (Garrappa, 2015) calculates Mittag-Leffler Function by inversion of Laplace transform.
The MittagLeffleR package (Gill & Straka, 2017) implements this algorithm and calculates pdf, cdf and random variates for both types of Mittag-Leffler distributions.
But this is more costly than method of log-moments.
Suppose now that \(\mathbf P(W_k > t) = L(t) t^{-\beta}\), where \(1 < \beta < 2\) and \(L(t)\) is a slowly varying function.
Convergence to b homogeneousb dynamics takes very long:
\[\mathbf E[N(t)] \approx \frac{t}{\mathbf E[W_1]} + \frac{C}{\mathbf E[W_1]^3} t^{2-\beta}, \quad \text{large } t\]
If the tail parameter satisfies \(1<\beta<2\), then \(\mathbf E[W_1] < \infty\), and in the limit as \(\ell \uparrow x_R\),
\[p_\ell T(\ell) = p_\ell \sum_{k=1}^{N_{p_\ell}} W_k \to {\rm Exp}(\mu_W).\]
Hence we b couldb approximate
\[T(\ell) \sim {\rm Exp}(\mu_W/p_\ell)\]
But we expect this convergence to be very slow, as seen in previous slide!
\[ \begin{multline} \sum_{k=1}^{N_{p_\ell}} W_k = \sum_{k=1}^{N_{p_\ell}} (W_k - \mu_W) + N_{p_\ell} \mu_W \\ = p_\ell^{-1/\beta} p_\ell^{1/\beta} \sum_{k=1}^{N_{p_\ell}} (W_k - \mu_W) + p_\ell^{-1}p_\ell N_{p_\ell} \mu_W \approx p_\ell^{-1/\beta} \sigma Z^{1/\beta} X + p_\ell^{-1} \mu_W Z \end{multline} \]
where \(Z \sim {\rm Exp}(1)\) and \(X \sim S(\beta, 1, 1, 0)\). The latter r.v. is geometric stable (Kozubowski & Rachev, 1999), and so:
\[T(\ell) = \sum_{k=1}^{N_{p_\ell}} W_k \approx {\rm GS}(\beta, 1, \sigma p_\ell^{-1/\beta}, p_\ell^{-1} \mu_W) \text{ for large } \ell. \]
Available estimators for \({\rm GS}\) distributions:
tempered heavy-tails
coupled CTRMs
Many datasets from finance (CGMY), geology and hydrology display tempered heavy tails:
This is modelled with one extra parameter by the tempered stable distribution
\[f_{\beta, \lambda}(x) = e^{-\lambda x + \lambda^\beta} f_\beta(x)\]
(where \(f_\beta(x)\) denotes the stable distribution.)
How to do inference for CTRMs with tempered heavy-tailed waiting times?