E-mail: B.Beranger@unsw.edu.au22affiliationtext: Department of Decisions Sciences, Bocconi University
E-mail: Simone.Padoan@unibocconi.it
ExtremalDep: Modelling extremal dependence in high-dimensional extremes
Abstract
From environmental sciences to finance, there are growing needs for assessing the risk of more extreme events than those observed. Extrapolating extreme events beyond the range of the data is not obvious and requires advanced tools based on extreme value theory. Furthermore, the complexity of risk assessments often requires the inclusion of multiple variables. Extreme value theory provides very important tools for the analysis of multivariate or spatial extreme events, but these are not easily accessible to professionals without appropriate expertise. This article provides a minimal background on multivariate and spatial extremes and gives simple yet thorough instructions to analyse high-dimensional extremes using the R package ExtremalDep. After briefly introducing the statistical methodologies, we focus on road testing the package’s toolbox through several real-world applications.
Keywords: Angular measure, Air pollution, Environmental data analysis, Exchange rates, Extremal coefficient, Extreme sets, Heat Waves, Max-stable processes, Multivariate generalized extreme-value distribution, Pickands dependence function, Quantile regions.
1 Introduction
In many applied fields, ranging from environmental sciences, finance to insurance, it is crucial to be able to assess the risk of more extreme events than those observed, with the goal to assess future catastrophes (e.g., a global financial crisis or large monetary losses due to natural hazards). The aim is ultimately to disclose the uncertainty about such extreme events to decision makers in order to carefully plan mitigation strategies. Extrapolating to extreme events lying beyond the range of existing data observations, is of high importance in risk management. The solution to this extreme value problem is not obvious since the occurrence probabilities () in the extrapolation regime are likely to be smaller than and therefore none of the observed data points exceed the desired event. Extreme Value Theory (EVT) provides a framework to develop probabilistic models and statistical methods for modelling extreme events. For a basic introduction, see for example beirlant2004, de2006 and falk2010.
In the univariate case, the theory and practice is mature with a significant number of packages available on the Comprehensive R Archive Network (CRAN), forming an extensive set of tools. Examples include ismev (ismev), evd Stephenson2002, evdbayes (evdbayes), extRemes (extRemes) and ExtremeRisks (extremerisks) to name a few. In practice, however, multiple variables are often observed. For instance, rainfall, temperature and wind may be observed at several locations spread across a region and contribute to the economic losses due to bad weather conditions; high pollution level is the result of a combination of multiple pollutants exceeding safety thresholds, etc. Focusing on multivariate and spatial problems, statistical methodologies for the analysis of high-dimensional extremes and their software implementation are less developed. The extremal dependence refers to the dependence structure behind multivariate extreme value distributions or the finite-dimensional distribution of stochastic processes for extremes. It is an infinite-dimensional (nonparametric) object subject to restrictive conditions, resulting in a complex formulation of the overall multivariate distribution and inferential challenges. The literature on high-dimensional extremes covers several topics including multivariate extremes (e.g. Beranger2015), spatial extremes (e.g davison2012statistical; davison2018), graphical models (graphical2020) and time series (e.g. Extremogram), and is remains nowadays a vibrating research field with a number of ongoing new results. These have produced useful R packages such as ExtremalDep (extremaldep), mev (mev), SpatialExtremes (SpatialExtremes210), mvPot (mvpot), RandomFields, graphicalExtremes graphicalextremes and extremogram (extremogrampkg). A review and comparison of statistical software for modelling extremes is presented in Belzile2023.
Most of the existing packages designed to analyse high-dimensional extremes allow for a parametric representation of the extremal dependence in order to handle it in a mathematically simpler way. ExtremalDep does the same thing, but also offers to model the extremal dependence in nonparametric and semiparametric ways, remaining more faithful to the nonparametric nature inherent to EVT. This manuscript illustrates the practical use of the package on important problems. In particular, it describes how to infer the extremal dependence in arbitrary dimension in a multivariate and spatial context, and how to use it to infer the probability of falling into a region considered extreme. Since the ordering between points is arbitrary in -dimensional Euclidean spaces, definitions of useful concepts including extreme sets, multivariate extreme quantiles, quantile regions, joint tail and conditional probabilities and multivariate return levels are provided. In practice, these are used to: analyze the risk of urban cities being subject to future episodes of high pollution; identify regions of a country that are subject to high risk of heavy precipitations events; evaluate the risk of some regions being subject to future heat wave episodes, and so on.
The article is divided into two parts, Section 2 discusses multivariate extremes while Section 3 focuses on spatial extremes. In the multivariate part, Section 2.1 gives a brief introduction to the theoretical background while Section 2.3 details how to perform parametric (frequentist and Bayesian) inference for the extremal dependence in arbitrary dimension and for joint and conditional tail probabilities. Section 2.4 explains how to perform nonparametric Bayesian inference for the extremal dependence in arbitrary dimension and, once the extremal dependence has been estimated, how to infer joint and conditional tail probabilities in the bivariate case. The focus then successively shifts to the random generation of bivariate extreme values with a flexible semi-parametric dependence structure, the estimation of small probabilities of belonging to certain extreme sets and the estimation extreme quantile regions corresponding to small joint tail probabilities.
2 Multivariate extremes
2.1 Background
Let be independent and identically distributed (i.i.d.) random vectors, where with , whose joint distribution is in the domain of attraction of a multivariate Generalised Extreme Value (GEV) distribution (de2006, chap. 6), shortly denoted as . This implies that for , there are norming sequences and such that
(2.1) |
where . The limiting distribution can be written as
(2.2) |
where is an extreme-value copula (beirlant2004, chap. 8) and for all the marginal distributions are of the form
(2.3) |
i.e., univariate GEV distributions with extreme value index describing the tail heaviness of the distribution (e.g., de2006, chap. 1). In particular, the extreme-value copula takes the form
(2.4) |
where is a homogeneous function of order named stable-tail dependence function (beirlant2004, chap. 8). This function is fully characterised by its restriction on which is known as the Pickands dependence function (e.g., falk2010, Ch. 4) and defined as
(2.5) |
where is a probability measure on the -dimensional unit simplex , named angular measure (e.g., falk2010, Ch. 4). A valid angular measure is a probability measure that satisfies the mean constraint (C1), while a valid Pickands dependence function is a function that must satisfies the convexity and boundary conditions (C2)-(C3). Specifically,
-
(C1)
.
-
(C2)
,
-
(C3)
.
Conditions (C2)-(C3) are necessary and sufficient to define a valid Pickands dependence function only in the bivariate case, (for details see beirlant2004, chap. 8). Since and are related through the one-to-one map in (2.5), the dependence parameter of the copula is commonly meant to be either the angular measure or the Pickands dependence function. Both are possible means to describe the so-called extremal dependence.
Given a sample with joint distribution , for each margin , maxima are computed over blocks of size , such that , where the th maximum given by , . The vector of componentwise maxima is then defined as . Setting , , , by (2.1), the joint distribution of can be approximated, for large enough , as where and each one-dimensional marginal distribution regarding the maximum can be approximated as
(2.6) |
The dependence among the maxima can be approximated by the extreme-value copula , where if the parameter is the angular measure , it describes the dependence among maxima as follows: the more mass places around the centre of the simplex the more dependent the maxima are. Conversely, the maxima becomes less dependent as concentrates its mass close to the vertices of the simplex. Alternatively, can be the Pickands dependence function with the lower and upper bounds in the inequality in (C3) respectively representing the cases of complete dependence and independence. A useful summary of the dependence among the maxima is given by the so-called extremal coefficient (e.g., beirlant2004, Ch. 8), defined as
(2.7) |
which represents the (fractional) number of independent block maxima. The extremal coefficient satisfies , with the lower and upper bounds describing the case of complete dependence and independence.
Beyond these results, the domain of attraction provides also a useful basis to derive very small joint tail probabilities corresponding to high-dimensional extremes relative to the original distribution or to derive extreme quantile regions corresponding to very small joint tail probabilities. These results are summarized as follows. The limit in (2.1) holds if and only if as , for all , with the convention that . Together with the weak convergence of the copula of the vector of componentwise maxima, namely as , for all , this result implies that for all , as ,
Recall that a copula defined on and associated to a joint distribution function , is a suitable function that allows the definition (see e.g., joe2014dependence, for details). For simplicity we assume that the margins of are continuous. The result in the above display implies the following important approximation. Let for and be the ()-quantile of , then by homogeneity of the stable tail dependence function, the probability that at least one among the variables exceeds a high quantile of its own distribution, can be approximated for large as
(2.8) |
By the inclusion and exclusion principle, the weak convergence result of the extreme value copula also allows to obtain
where is known as the tail copula function, which is also a homogeneous function of order , and therefore the probability that all the variables exceed simultaneously high quantiles of their own distributions can be approximated, for a sufficiently large , as
(2.9) |
Finally, a related question is, how to determine an extreme region given a small joint probability of falling in it? The approach discussed by cai2011, einmahl2013 and he2017 is summarized as follows with a focus on the bivariate case explored by einmahl2013. Let be the density on of a bivariate distribution satisfying , with. Define quantile regions by the level sets of such that
(2.10) |
with . Then, for a small probability we derive the region such that and everywhere on is smaller than everywhere on and therefore the latter is the smallest region satisfying . Accordingly, an extreme quantile region is defined by a level set (with in (2.10)), which is such that where and the expected number of points falling in it is as . Note that when , the region is so extreme that almost no observations are expected to fall in it.
The domain of attraction condition implies the existence of a measure , named exponent measure, such that
and satisfying , for all Borel set that are bounded away from the origin. Assume that is bounded away from zero on and, outside this region, is decreasing in each coordinate, for some . Assume also that the exponent measure admits a density function such that
and
In particular, we have that is related to the density of the angular measure by , where and . The basic set is defined as , where
and, according to the exponent measure, its size is equal to
The set is then suitably transformed in order to obtain for the approximation
(2.11) |
where with as and . Furthermore, and with , where is a suitable positive scale function (de2006, Ch. 1-2). is a good approximation of in the sense that and as (see einmahl2013), where .
2.2 Statistical challenges
From a statistical perspective, the multivariate GEV model defines a complex semiparametric family of distributions of the form , , where is a finite-dimensional vector of marginal parameters and is an infinite-dimensional dependence parameter of the copula , named extremal dependence, living in a suitable space, depending whether is the angular measure or the Pickands dependence function. The estimation of such infinite-dimensional parameter, accounting for the corresponding constraints (see conditions (C1)-(C3) of Section 2.1), is far from being a simple task. To reduce complexity, classes of parametric models were initially proposed to model , see Beranger2015 for a review. In order to provide valid extremal dependence structures, the proposed models satisfy the required constraints. Section 2.3 describes how to fit some of the most popular parametric models for using ExtremalDep.
Over time, more sophisticated semi- and non-parametric approaches for modelling and estimating the extremal dependence have been proposed, under both parametrisations. Some examples include: projection estimators (e.g., fils2008), polynomials (e.g., Marcon2017a) and splines (e.g., cormier2014), just to name a few. Section 2.4 details how to use ExtremalDep for non-parametric estimation of the extremal dependence, simulation of bivariate extreme values with a flexible semi-parametric dependence structure, estimation of small probabilities to belong to certain extreme sets and estimation extreme quantile regions corresponding to small joint tail probabilities.
2.3 Parametric modelling of the extremal dependence
A non-exhaustive list of popular parametric models for the extremal dependence includes the Asymmetric-Logistic, Pairwise-Beta, Tilted-Dirichlet, Hüsler-Reiss, Extremal- and Extremal-Skew- models (see e.g., Beranger2015; Beranger2017).
The Poisson-Point-Process (PPP) is a simple estimation method for statistical models used to represent the extremal dependence, when the latter is parametrized according to the angular measure and provided that it allows for a density on , where is a suitable vector of parameters (see e.g., coles1991; engelke2015). In particular, let be i.i.d. observations from and be their equivalent with unit Fréchet margins. The pseudopolar transformation is defined as where is the radial part and is the angular part of , . The domain of attraction condition implies that if is a large threshold and is a set of points with a radial component larger than , then the number of points falling in is described approximatively by a Poisson random variable with rate , where . Thus, conditionally on observing points with radial part greater than , such points are approximately independent and generated from the density function . The likelihood function of the exceedances can then be approximated as
Estimates of can be obtained by maximizing the log-likelihood . Alternatively, one can appeal to a Bayesian approach (sabourin2013), based on the approximate posterior density of the angular density parameters given by
where is a suitable prior distribution on and is a suitable parameter space.
The routine fExtDep with argument method="PPP" allows to estimate by likelihood maximisation. The model class for is specified with the argument model with the list of available options reported in Table 2.3. Some classes of models are characterised by an angular density that places a continuous mass only in the interior of (PB, TD, HR), while others also allow to place densities and atoms on subspaces of . Beranger2017 describe an extended version of the ML method that allows for the estimation of these more complex models up to the three-dimensional case (AL, ET, EST) by exploiting the following approach. In the bivariate case, an observation falls at an endpoint of if or for some small and on the interior of otherwise. A trivariate observation falls: at the th corner of if , on the edge between the th and th components if , and in the interior of otherwise (i.e., if for all ) (see Beranger2017, for details). This approach is implemented when specifying a positive value for the argument c of fExtDep.
We now illustrate the use of fExtDep on the analysis of high levels of air pollution recorded in Leeds, UK, over the winter period November 1st to February 27/28th, between 1994 and 1998 (see heffernan2004). The object pollution contains the daily maximum of the five air pollutants: particulate matter (PM10), nitrogen oxide (NO), nitrogen dioxide (NO2), ozone (03), and sulfur dioxide (SO2), in unit Fréchet scale (for a description on the used transformation see Beranger2015). Furthermore, the datasets PNS, NSN and PNN contains the angular components corresponding to the 100 largest radial components of the triplets of pollutants (PM10, NO, SO2), (NO2, SO2, NO) and (PM10, NO, NO2), respectively. We run the following simple commands
R> data(pollution)R> f.et05 <- fExtDep(method="PPP", data=PNS, model = "ET",+ par.start = c(0.5, 0.5, 0.5, 3), trace=2, c=0.05)
initial value -118.269640iter 10 value -146.907539iter 10 value -146.907539final value -146.907539converged
R> f.et <- fExtDep(method="PPP", data=PNS, model = "ET",+ par.start = c(0.5, 0.5, 0.5, 3), trace=2 )
initial value -161.078869iter 10 value -225.654851iter 20 value -234.524086iter 30 value -235.285461iter 40 value -235.360084final value -235.382293converged The first call to the fExtDep routine fits the Extremal- angular density (model="ET"), to the pollution data (data=PSN), allowing for point masses at the corners by specifying the argument c, while the second call considers a density defined only in the interior of . The argument par.start gives starting values for the parameters and trace=2 allows to monitor optimization progress. The optimization method is set by the argument optim.method ("BFGS" by default) on which basis the routine optim from the stats library is used. fExtDep returns a list with the estimates of the model parameters (par), their standard errors (SE), the log-likelihood maximum (LL) and the Takeuchi Information Criterion (TIC). fExtDep is first ran for the extremal- model using different values of ( and ) to take into account point masses at the corners of . The TICs indicate that only assuming mass in the interior of provides the best results (lowest TIC). A performance comparison between the PB, AL, TD, HR, ET and EST models defined on the interior only, is performed with TIC scores reported in Table 1 and as a result the Extremal skew- model fits the data best (lowest TIC).
model | PB | AL | TD | HR | ET | EST |
---|---|---|---|---|---|---|
TIC | -188.75 | -406.85 | -393.75 | -460.25 | -461.74 | -493.31 |
Bayesian inference for the parameter is implemented through a model averaging approach (sabourin2013) and available in the routine fExtDep when method="BayesianPPP". Table 2.3 reports the implemented prior densities for each angular density model. In particular, stands for a one-dimensional Gaussian density with mean and variance , is a transformation applied to the parameter (to map it to the real line). For each component of , the proposal density is also Gaussian but of the form , where and are the proposed value and the value at the th iteration of the algorithm of the element of , respectively, and MCpar is the variance term that needs to be specified by the user. The next display illustrates how to perform Bayesian inference for the Hüssler-Reiss model (model="HR") considering Nsim=5e+4 iterations, a burn-in period of Nbin=3e+4 iterations, hyper-parameters Hpar=Hpar.hr, MCpar=0.35 and using the argument seed for reproducibility . Refer to the help page for a description of the list returned as output from fExtDep.
R> Hpar.hr <- list(mean.lambda=0, sd.lambda=3)R> PNS.hr <- fExtDep(method="BayesianPPP", data=PNS, model="HR", Nsim=5e+4,+ Nbin=3e+4, Hpar=Hpar.pb, MCpar=0.35, seed=14342)
|=====================================================================| 100%
R> labs <- c(expression(PM[10]),expression(NO),expression(SO[2]))R> plot.ExtDep(object="angular", model="HR", par=PNS.hr$emp.mean,+ data=PNS, cex.lab=2, labels=labs)
Model | model | Prior | ||
Pairwise-Beta | PB | ( ; mean.alpha, sd.alpha) | ||
( ; mean.beta, sd.beta) | ||||
\cdashline1-5 Asymmetric-Logistic | AL | |||
logit | ||||
\cdashline1-5 Tilted-Dirichlet | TD | |||
\cdashline1-5 Hüsler-Reiss | HR | |||
\cdashline1-5 Extremal- | ET | atanh | ||
; mean.nu, sd.nu) | ||||
\cdashline1-5 Extremal-Skew- | EST | atanh | ||
identity | ||||
The empirical mean (standard deviation) obtained from the approximate posterior distribution of are , while the Bayesian Information Criterion (BIC) is . Visualisation of the fitted angular density is obtained from the routine plot.ExtDep by specifying object = "angular" and model="PB", and selecting the parameters to be the mean of the approximate posterior distribution (par=fit.hr$emp.mean). The argument data = PNS plots the angular data on top of the estimated density. Figure 1 showcases a dependence structure that appears to follow the data structure well but other models should be fitted (see Beranger2015, for further details).
The purpose of estimating the extremal dependence is to estimate small tail probabilities as in (2.9). Similar to Cooley2010, we define the extreme event and infer this probability using the routine pExtDep with method="Parametric", type="upper" and specifying the argument model. Providing a vector to par yields a point estimate of the tail probabilities while a matrix (e.g., the posterior sample) with rows of parameter values derives an approximate posterior distribution for such probabilities. Note that the object est and the routine transform are used to transform values to unit Fréchet scale following the steps of heffernan2004.
R< est.fun <- function(x){+ x <- na.omit(x)+ unlist(evd::fpot(x, threshold=quantile(x, probs=0.7))+ [c("threshold", "estimate")])+ }R> est <- apply(winterdat, 2, est.fun)R> transform <- function (x, data, par){+ data <- na.omit(data)+ if(x > par[1]){+ emp.dist <- mean(data <= par[1])+ dist <- 1-(1-emp.dist )*max(0, 1+par[3]*(x-par[1])/par[2])^(-1/par[3])+ }else{+ dist <- mean(data <= x)+ }+ return(-1/log(dist))+ }R> th <- c(95, 270, 110, NA, 95)R> Th <- sapply(c(1:3,5), function(x) transform(th[x], data=winterdat[,x], par=est[,x]) )R> names(Th) <- colnames(winterdat[c(1:3,5)])R> xl.PNS <- bquote("P(" * PM[10] ~ ">" ~ .(th[1]) * ", NO" ~ ">" ~ .(th[2])+ * ", " * SO[2] ~ ">" ~ .(th[5]) * ")")R> P.PNS <- pExtDep(q=Th[colnames(PNS)], type="upper", method="Parametric",+ model="HR", par=PNS.hr$stored.vals, xlab=xl.PNS) The left panel of Figure 2 provides the approximate posterior distribution of the tail probabilities in (2.9) for the extreme events. Note that a smoother output could be obtained by considering Nsim=5e+5 and Nburn=3e+5. Formula (2.9) can also be used to derive a possible definition of the so-called joint return level, see Beranger2015 for details, i.e. the sequence of values , that satisfies the equation p=P(Y_j¿y_j;p, X_i¿x_i, j∈J, i∈{1,…,d}\J), for a given small probability , where , with is a sequence of fixed high thresholds. Below is an example where the routine plot_ExtDep is used to infer joint return levels. Here only a subset of the posterior is considered for a quick evaluation but in practice, we do recommend using the full posterior.
R> Q.fix <- c(NA, Th[c(2,4)])R> PM10.range <- seq(from=est[1,1], to=400, by=5)R> Q.range <- sapply(PM10.range, transform, data=winterdat[,1], par=est[,1])R> set.seed(1)R> ind <- sample(1:2e+4, 100)R> rl.PM10 <- plot_ExtDep(object="returns", model="HR", par=PNS.hr$stored.vals[ind,], Q.fix=Q.fix, Q.range=Q.range, Q.range0=PM10.range, labels=expression(PM[10]), main=bquote("Return level for" ~ PM[10] ~ "when NO" ~ ">" ~ .(th[2]) ~ "and" ~ SO[2] ~ ">" ~ .(th[5]) ), cex.lab=1.4, cex.axis=1.4) Returns levels are displayed when object="returns". In addition, it requires to provide Q.fix, a vector of length equal to the model dimension ( or ), where quantile values can be fixed for some components while others (NAs) are left to vary. The Q.range argument provides a vector (or matrix) of quantile values on the unit Fréchet scale, for those that aren’t fixed. If Q.fix contains a single NA then Q.range must be a vector or a single column matrix. Q.range0 provide the same sequences as Q.range but on the original scale. This plotting procedure relies on the pExtDep routine.
The middle panel of Figure 2 reports the univariate joint return level curve for jointly to the extreme event , corresponding to the return period , which has been estimated by the approximate posterior mean and uncertainty is given by the 95% (pointwise) credibility intervals. Similarly, the right panel of Figure 2 depicts the posterior mean and credibility intervals of the estimated contour levels of the bivariate return levels for jointly to extreme event . Note that conditional return levels can be obtained by specifying cond=TRUE, the conditional event being the fixed event.
2.4 Semi- and non-parametric modeling of the extremal dependence
Common objectives when analyzing extreme values include assessment of the dependence level among extremes, e.g. using the dependence structures in (2.4)-(2.5), simulation of multiple extremes, estimation of small joint tail probability, e.g. (2.8)-(2.9), and estimation of extreme quantile regions. The next sections describe the steps required to accomplish such goals.
2.4.1 Extremal dependence estimation via Bernstein polynomials
Below, we give a brief summary of a simple estimation method for the extremal dependence with an arbitrary number of componentwise maxima proposed by Marcon2017a. We refer to the aforementioned paper for further details.
The method consists of two steps: 1) a nonparametric pilot estimation of the Pickands dependence function, 2) a regularization of such estimate by projecting it into a polynomial representation in Bernstein form imposing conditions (C2)-(C3). Given some i.i.d. random vectors (approximately) distributed as the multivariate GEV distribution in (2.2), the first step is achieved by using the madogram-based estimator of the Pickands dependence function. This is given, for all , by ^A_k(t)=^νk(t)+d-1∑j=1d(tj/(1+tj))1-^νk(t)-d-1∑j=1d(tj/(1+tj)), where ^ν_k(t)=1k∑_i=1^k(max_1⩽j⩽d F_k,j^1/t_j(Y_i,j)-1d∑_j=1^dF_k,j^1/t_j(Y_i,j)), and denotes the empirical distribution of the th variable. For the regularization of the pilot estimate, take and let be the set of multi-indices such that and , whose cardinality is denoted by . Let
(2.12) |
be the Bernstein-Bézier polynomial representation of the Pickands dependence function, where for each ,
(2.13) |
is the Bernstein polynomial basis function of index and degree .
Now, let be the family of functions satisfying conditions (C2)-(C3), and be a sequence of families of constrained multivariate Bernstein–Bézier polynomials on , where is the row vector , is a column vector, is a suitable matrix of full row rank and is a vector. The constraint on the coefficient vector guarantees that each member of satisfies (C2)-(C3). A projection estimator of the Pickands dependence function based on the estimator is the solution to the following optimization problem ~A_k,κ=arg min_f∈A_κ∥ ^A_k - f∥. For a finite set of points , with and such a solution is obtained by finding the minimizer of the constrained least-squares problem ^β_κ=arg min_β_κ∈[0,1]^N_κ:R_κβ_κ⩾r_κ1U∑_u=1^U(b_κ(t_u)β_κ-^A_k(t_u))^2, which expression is ^β_κ=(b^⊤_κb_κ)^-1b_κ^⊤^A_k -(b^⊤_κb_κ)^-1r_κ^⊤λ, where is a vector of Lagrange multipliers.
This method is implemented in the routine beed of ExtremalDep and its usage is demonstrated through the analysis of heavy rainfall in France. Hydrologists are interested in identifying different geographic regions that differ from each other in that there are clusters of weather stations whose data exhibit substantially different levels of extreme dependence. Within a cluster, climate characteristics are expected to be homogeneous, whereas they can be quite heterogeneous between clusters. The dataset is available through the PrecipFrance object which consists of weekly maxima of hourly rainfall ($precip) recorded at weather stations in France, during the Fall season between 1993 and 2011, yielding a sample size of observations. Coordinates of each station are stored in the list elements $lat and $lon. Note that hourly rainfall has been appropriately pre-processed and with them the weekly maxima have been already computed from bernard2013clustering. By applying the algorithm proposed by bernard2013clustering to the weekly maxima of hourly rainfall, the weather stations are divided into clusters. This is obtained using the PAMfmado routine through the following commands
R> data(PrecipFrance)R> attach(PrecipFrance)R> nclust <- 7R> PAMmado <- PAMfmado(precip,nclust) For each cluster, stations are randomly selected in order to have equal size () clusters and for each group, the Pickands dependence function is estimated using the Bernstein projection estimator based on the madogram with polynomial degree equal . To summarise, an estimate of the extremal coefficient is computed using (2.7), i.e. , through the following commands
R> clust <- PAMmado$clusteringR> d <- 5R> stationsn <- matrix(NA, nclust, d)R> xx <- simplex(d=d, n=15)R> fit <- list(length=nclust)R> est <- vector(length=nclust)R> set.seed(1)R> for(i in 1:nclust){+ stationsn[i,] <- sample(which(clust==i), 5)+ data_tmp <- precip[,stationsn[i,]]+ data_uf_tmp <- trans2UFrechet(data_tmp, type="Empirical")+ fit[[i]] <- beed(data_uf_tmp, xx, d, "md", "emp", k=7)+ est[i] <- fit[[i]]$extind+ } The left panel of Figure 3 indicates the extremal dependence is strongest in the center of the country, away from the coasts, where the conjunction of different densities of air masses produces extreme rain storms. This is consistent with what climatologists expect since extreme precipitation that affects the Mediterranean coast in the fall is caused by the interaction of southern and mountain winds coming from the Pyrénées, Cévennes and Alps regions. In the north, heavy rainfall is produced by mid-latitude perturbations in Brittany or regions further north and Paris. Within clusters, extremes are strongly dependent. The right panel of Figure 3 shows the pairwise extremal coefficients from all stations, computed through the estimated Pickands dependence functions using the raw madogram estimator (MD) and its Berntein projection (MD-BP), versus the geodesic distance between sites. We have for the locations that are less than 200 km apart, meaning that extremes are either strongly or mildly dependent, while for sites more than 200 km apart, we have , meaning that extremes are at most weakly dependent or even independent. The graph also shows the benefits of the projection method: after projection, the extremal coefficients fall within the admissible range . The beed function is used for the MD-PB estimator whereas madrogram implements the raw madogram estimator MD. The following set of commands were used for the estimation.
R> library(geosphere)R> pairs <- t(combn(92,2))R> pairs.LLO <- cbind(pairs, lon[pairs[,1]], lat[pairs[,1]],+ lon[pairs[,2]], lat[pairs[,2]])R> pairs.dist <- apply(pairs.LLO, 1, function(x) distm(x=x[3:4], y=x[5:6],+ fun=distGeo) )/1000R> pairs.EC <- pairs.ECBP <- vector(length=nrow(pairs))R> S <- simplex(d=2, n=49)R> for(i in 1:nrow(pairs)){+ data.tmp <- trans2UFrechet(precip[,pairs[i,]], type="Empirical")+ fit.tmp <- beed(data.tmp, S, 2, "md", "emp", k=7)+ pairs.ECBP[i] <- fit.tmp$extind+ pairs.EC[i] <- 2*madogram(S, data.tmp, "emp")[which(S[,1]==0.5)]+ }


2.4.2 Bernstein polynomials modeling and Bayesian nonparametric inference
In many applications it is crucial to estimate joint probabilities such as , for pair of extreme values value which allows the estimation of related quantities such as small conditional probabilities with . Section 2.1 describes a theoretical framework that can be used to approximate such probabilities and we now briefly describe a Bayesian nonparametric inference method based on Bernstein polynomials introduced by marcon2016.
Let be (independent) observations (approximately) distributed as a bivariate GEV distribution. First, a prior distribution on the parameters , of the marginal GEV distributions is specified as with , i.e. a product of uniform prior distributions on the real line for , and . Note that this improper prior distribution leads to a proper posterior distribution (northrop2016). Samples from the posterior distribution are generated using an adaptive (Gaussian) random-walk Metropolis-Hastings (RWMH) algorithm (haario+st01; garthwaite2016). At the current state of the chain, is updated by the proposal , where is a -dimensional Gaussian density with mean and covariance . The proposal covariance matrix is specified as
(2.14) |
where is the -dimensional identity matrix, , and is a scaling parameter that affects the acceptance rate of proposal parameter values (haario+st01) and, is updated using a Robbins-Monro process so that
(2.15) |
where is a steplength constant, , and is the univariate standard Gaussian distribution function (garthwaite2016). To control the desired sampler acceptance probability, the parameter is specified as in roberts1997.
In the bivariate case, the polynomial Pickands dependence function in Bernstein form becomes
for , where and the polynomial basis in (2.13) reduces to b_j(t; κ) =κ!j! (κ-j)! j^j(1-j)^κ-j, j=0,…,κ. For fixed degree , marcon2016 derived the restrictions on so that satisfies conditions (C2)-(C3) and is therefore a valid Pickands dependence function. They also demonstrated that the polynomial implies that the distribution of the corresponding angular measure can be written as a Bernstein polynomial of degree ,
(2.16) |
where . Additionally, marcon2016 established the conditions on the coefficients so that is the distribution of a valid angular measure and proposed a Bayesian nonparametric procedure for the inference of both and , which can then be used to compute an approximation of joint tail probabilities. The proposed method consists of three main steps: 1) the specification of a prior distribution for the polynomial order and coefficients , 2) the derivation of the likelihood function, 3) the definition of a Markov Chain Monte Carlo (MCMC) algorithm for the posterior distribution computation. In particular, , where is a prior on the polynomial order (e.g., Poisson, negative Binomial, etc.) and is a prior on the coefficients , where and are the priors on the coefficients representative of the atoms and at the edges of the interval . Such priors are specified as uniform distributions on suitable intervals, which have been chosen to ensure that the resulting Bernstein polynomial satisfies the constraint (C1). Specification of the prior induces also a prior on the coefficients of the corresponding polynomial , which automatically satisfy constraints (C2)-(C3). To deal with the fact at each MCMC iteration the dimension of changes with , a trans-dimensional MCMC scheme is considered following marcon2016. At the state , is updated using the proposal distribution , where is defined such that if , it places mass on with probability and if it places mass on and with equal probability. Finally, the likelihood function is defined as , where and for any such that with ,
and where , , , see marcon2016 and beranger2021 for details. The MCMC scheme for the joint inference of marginal distribution and extremal dependence is reported in Algorithm 1.
For pairs of future unobserved yet extremes values, the joint exceeding probability can be estimated through the Bayesian paradigm using the posterior predictive distribution which can be approximated, given a sample with , from the posterior distribution as
(2.17) | ||||
where are distributed as a bivariate GEV with unit-Fréchet margins, with and , , denotes the distribution function of a Beta random variable with shape parameters (see marcon2016, for details).
We show the utility of the methodology by analyzing the joint extremal behavior of log-returns of exchange rates between Pound Sterling and U.S. Dollar (GBP/USD), and Pound Sterling and Japanese Yen (GBP/JPY). The data is available as logReturns and consists of the monthly maxima of daily log-returns exchange rates from March 1991 to December 2014 ( observations). Exchange rates are $USD and $JPY, while $date_USD and $date_JPY are the date when the monthly maxima was attained. First the trend and seasonality are removed from each maxima series using the ts() and stl() functions from the stats package.
R> data(logReturns)R> mm_gbp_usd <- ts(logReturns$USD, start=c(1991,3), end=c(2014,12), frequency=12)R> mm_gbp_jpy <- ts(logReturns$JPY, start=c(1991,3), end=c(2014,12), frequency=12)R> seas_usd <- stl(mm_gbp_usd, s.window="period")R> seas_jpy <- stl(mm_gbp_jpy, s.window="period")R> mm_gbp_usd_filt <- mm_gbp_usd - rowSums(seas_usd$time.series[,-3])R> mm_gbp_jpy_filt <- mm_gbp_jpy - rowSums(seas_jpy$time.series[,-3])


The top-left and bottom-left panels of Figure 4 display the de-trended and de-seasonalised maxima and the right panel scatter plot shows strong dependence between the extremes of both exchanges rates. The extremal dependence structure is estimated using the Bayesian nonparametric framework described above. The fExtDep.np routine allows for nonparametric estimation of the extremal dependence and method = "Bayesian" specifies that such a dependence is in Bernstein polynomial form and a Bayesian approach is used for the inference. The argument mar.fit = TRUE (default) allows for joint estimation of the margins and dependence while mar.prelim = TRUE (default) fits the marginal distributions using the RWMH algorithm and the fGEV routine, in order to obtain starting values for the marginal parameters. The prior distribution for is set to be a negative binomial on with mean and variance , and for the point masses and uniform distributions on and respectively, where and . Below, only the hyper-parameters need to be specified since prior.k="nbinom" and prior.pm="unif" by default. Lastly, a two-column object mm_gbp representing the data is created.
R> hyperparam <- list(mu.nbinom = 3.2, var.nbinom = 4.48)R> mm_gbp <- cbind(as.vector(mm_gbp_usd_filt), as.vector(mm_gbp_jpy_filt))R> set.seed(123)R> gbp_mar <- fExtDep.np(method="Bayesian", data=mm_gbp,+ par10=rep(0.1, 3), par20=rep(0.1,3),+ sig10=0.0001, sig20=0.0001, k0=5,+ hyperparam = hyperparam, nsim=5e+4)
Preliminary on margin 1|=====================================================================| 100%Preliminary on margin 2|=====================================================================| 100%Estimation of the extremal dependence and margins|=====================================================================| 100%
R> diagnostics(gbp_mar) The gbp_mar object contains the posterior samples for all the parameters: point mass and , polynomial coefficients and degree , marginal parameters and , respectively reported by the arguments pm, eta, k, mar1 and mar2. Its also contains binary vectors indicating the accepted marginal and dependence proposals (accepted.mar1, accepted.mar2 and accepted) and the marginal proposals that were rejected right away for not being in the parameter space (straight.reject1 and straight.reject2). It also includes acceptance probabilities at each step (acc.vec, acc.vec.mar1 and acc.vec.mar2), the marginal scaling parameters and (sig1.vec and sig2.vec), as well as some of the inputs. The diagnostics function investigates the convergence of the algorithm, producing Figure 5. The top panels display the scaling parameters and from the marginal proposals and the polynomial degree , as function of the iterations. The bottom panels show the corresponding acceptance probability with the desired acceptance probability of (horizontal black solid lines). Overall, Figure 5 suggests a burn-in period of iterations.

The summary.ExtDep routine computes summary statistics of the MCMC output, including posterior sample, mean and credibility intervals (can be modified using argument cred) for the angular density and Pickands dependence function. Posterior mean and credibility intervals are also computed for all parameters. Setting plot=TRUE displays the posterior mean and credibility intervals for both angular density and Pickands dependence function, as well as the prior and posterior distribution for the point mass and the polynomial degree . The returns routine inputs the outputs of the fExtDep.np and summary.ExtDep (via the arguments out and summary.mcmc) to compute exceeding probabilities as defined in (2.17), for extreme values specified by y. The argument plot allows to visualize such probabilities as long as y defines a square grid and data adds the relevant datapoints. Usage of the summary.ExtDep and returns functions is provided below with graphical outputs presented in Figure 6.
R> gbp_mar_sum <- summary.ExtDep(mcmc=gbp_mar, burn=30000, plot=TRUE)R> mm_gbp_range <- apply(mm_gbp,2,quantile,c(0.9,0.995))R> y_gbp_usd <- seq(from=mm_gbp_range[1,1], to=mm_gbp_range[2,1], length=20)R> y_gbp_jpy <- seq(from=mm_gbp_range[1,2], to=mm_gbp_range[2,2], length=20)R> y <- as.matrix(expand.grid(y_gbp_usd, y_gbp_jpy, KEEP.OUT.ATTRS = FALSE))R> ret_marg <- returns(out=gbp_mar, summary.mcmc=gbp_mar_sum, y=y,+ plot=TRUE, data=mm_gbp,+ labels=c("GBP/USD exchange rate", "GBP/JPY exchange rate"))


As mentioned at the beginning of this section, computing small conditional probabilities such as or can be of interest. To proceed, take which corresponds to the observed quantiles. The joint probability of exceedance is computed (through returns) using the approximation , where and are the transformation of and to unit Fréchet scale. Using the pGEV function to evaluate the marginal probability of exceedance, we obtain and . The computations are presented in the code below.
R> qs <- apply(mm_gbp,2,quantile,c(0.99))R> jointP <- returns(out=gbp_mar, summary.mcmc=gbp_mar_sum,+ y= matrix(qs, ncol=2) )par1 <- gbp_mar_sum$mar1.meanpar2 <- gbp_mar_sum$mar2.meanjointP / pGEV(qs[1], loc=par1[1], scale=par1[2], shape=par1[3], lower.tail=F)jointP / pGEV(qs[2], loc=par2[1], scale=par2[2], shape=par2[3], lower.tail=F)
2.4.3 Simulation of extreme values
In environmental statistics, there is active research in “stochastic weather generators” which aims at simulating realisations from atmospheric variables according to some stochastic representation. A challenging task is the simulation of high-dimensional extremes, since their extremal dependence (Pickands dependence function or angular measure) is an infinite dimensional parameter of the multivariate GEV distribution and is therefore challenging to estimate. The ability to simulate high-dimensional extremes permits to approximate the tail probabilities in (2.8) and (2.9), even though few extremes are available in the dataset.
For simplicity, most of the simulation methods assume that the extremal dependence complies with some parameter model, while few attempt to consider a nonparametric approach. Marcon2017 proposed a flexible procedure for sampling from bivariate extremes with a semiparametric dependence structure, which is summarized as follows. For every small probabilities and such that and , the tail probabilities in (2.8) and (2.9) can be approximated by and . For such probabilities and , we have that and , where is a random variable on distributed according to the angular distribution . Now, let , where is a unit-Pareto random variable, then and . Marcon2017 proposed to model through Bernstein polynomials and demonstrated that can be written as a finite mixture of Beta distributions with weights defined by a suitable transformation of the polynomial coefficients. This led to a simple algorithm to sample from (see Algorithm 1) and an algorithm for sampling observations from the tail of a bivariate distribution (see Algorithm 3), which can be used to approximate the corresponding tail probabilities. Briefly, from (2.6) we have that for a sufficiently large and a sufficiently small (with as ), , with . Let and be two high thresholds such that for every and for which the above approximations hold, as well as and , we have , . For the failure regions and we have and , where , for . By simulating a large sample of angular components from and radial components from a unit-Pareto distribution, we compute , for and estimate of the probability of falling in and by
(2.18) |
The methodology is illustrated on the Parcay-Meslay dataset which consists of daily maxima of hourly wind speed (WS) and wind gust (WG) in meters per second (m/s) and differential of daily range of the hourly air pressure (DP) at sea level in millibars. Measurements are taken in the city of Parçay-Meslay, located in the northwest of France, from July 2004 to July 2013. For this analysis, we focus on positive values of DP implying an increase in the daily air pressure level. High air pressure levels are associated with a high content of water vapor in the air which often occurs in stormy weather and leads to strong winds. For brevity, we focus on the relationship between WS and DP. The GEV parameters are first estimated using the Point Process approach (e.g. Coles2001, Ch. 7), implemented in the fpot routine of the evd package (Stephenson2002), where the observed marginal quantiles are set as suitable thresholds. The margins are transformed to unit-Fréchet scale using the trans2UFrechet routine with argument type="empirical", meaning the transformation is applied, where denotes the empirical distribution of the -th component.
R> data(WindSpeedGust)R> years<- format(ParcayMeslay$time, format="%Y")R> attach(ParcayMeslay[which(years %in% c(2004:2013)),])R> WS_th <- quantile(WS,.9)R> DP_th <- quantile(DP,.9)R> pars.WS <- evd::fpot(WS, WS_th, model="pp")$estimateR> pars.DP <- evd::fpot(DP, DP_th, model="pp")$estimateR> data_uf <- trans2UFrechet(cbind(WS,DP), type="Empirical") The angular distribution is then estimated using the approximate likelihood approach (see Beranger2015). Observations with radius component greater than their quantile are selected and the routine fExtDep.np is then called with arguments method="Frequentist" and type="maxima" to maximize the likelihood of a polynomial angular distribution in Bernstein form through the non-linear optimization routine nloptr from the nloptr package (nloptr) subject to constraints established in Marcon2017a. Empirical studies (Marcon2017a) suggest that the polynomial degree is enough. When type="rawdata, data are extracted using a threshold on the radial component set by the argument u and the likelihood for a sample maxima written as function of the Pickands dependence function in Bernstein form is optimized. In addition, when mar.fit=TRUE then marginal empirical transformation to unit Fréchet of the data is applied. The plot.ExtDep routine displays graphical summaries (Pickands dependence function and angular density) of the estimated dependence structure (see left and middle panels of Figure 7). A moderate level of dependence is observed as well as point masses at the vertices. The routine rExtDep with arguments model="semi.bvevd" and angular=TRUE generates pseudo-angles according to Algorithm 1 of Marcon2017. The right panel of Figure 7 presents a histogram of randomly generated angles and point-masses (black triangles), the red line and dots represent the estimated dependence structure.
R> rdata <- rowSums(data_uf)R> r0 <- quantile(rdata, probs=.90)R> extdata <- data_uf[rdata>=r0,]R > SP_mle <- fExtDep.np(method="Frequentist", data=extdata, k0=10, type="maxima")R> plot.ExtDep.np(out=SP_mle, type="summary")R> SP_wsim <- rExtDep(nsim, model="semi.bvevd", param=SP_mle$Ahat$beta,+ angular=TRUE)



The routine rExtDep performs random generation of bivariate maxima (model="semi.bvevd") and bivariate exceedances (model="semi.bvexceed"), according to Algorithm 2 and 3 of Marcon2017. When angular=TRUE solely angular components are generated, no matter the model argument, meaning that setting model="semi.bvexceed" in the above call of rExtDep would produce the same result. The argument mar allows for transformations to GEV margins. When model="semi.bvexceed", one can choose to simulate bivariate observations from the failure regions (exceed.type="or") or (exceed.type="and"), where is a suitable threshold specified by threshold. The code below generates observations from both failure regions with threshold , above the marginal quantiles (see Figure 8 for an illustration).
R> set.seed(10)R> SP_exceed_or <- rExtDep(n=200, model="semi.bvexceed", param=SP_mle$Ahat$beta,+ mar=rbind(pars.WS, pars.DP), threshold=c(10,20),+ exceed.type="or")R> SP_exceed_and <- rExtDep(n=200, model="semi.bvexceed", param=SP_mle$Ahat$beta,+ mar=rbind(pars.WS, pars.DP), threshold=c(10,20),+ exceed.type="and")

Finally, the routine pFailure computes empirical estimates of probabilities of belonging to the failure sets (type="or") and (type="and) by applying formula (2.18) on data generated by rExtDep. Both probabilities are computed when type="both". The argument plot offers the possibility to display contour plots when sequences of thresholds are provided through the arguments u1 and u2. The following code generates samples to estimate the probabilities of belonging to each failure sets for thresholds ranging from 19 to 28 for the wind speed and from 40 to 60 for the differential of pressure with outputs given in Figure 9.
R> pF <- pFailure(n=50000, beta=SP_mle$Ahat$beta,+ u1=seq(from=19, to=28, length=200), mar1=pars.WS,+ u2=seq(from=40, to=60, length=200), mar2=pars.DP, type="both",+ plot=TRUE, xlab="Daily-maximum Wind Speed (m/s)",+ ylab="Differential of Pressure (mbar)", nlevels=15)


2.4.4 Estimation of extreme quantile regions
As highlighted in Section 2.1, an important problem is to define a quantile region as in (2.10), given a very small probability to fall in it. Given a sample of size , an extreme quantile region can be defined by the level set such that , where satisfies and as . For practical purposes, can be approximated by the region given in (2.11) which requires estimating the marginal parameters , , the basic set and the density of the angular measure, allowing, in turn, to obtain an estimate of . beranger2021 discussed a Bayesian approach for the inference of based on a polynomial angular measure in Bernstein form and the censored-likelihood approach (e.g. beirlant2004, Ch. 8). Since such a methodology is similar to the one described in Section “Bernstein polynomials modeling and Bayesian nonparametric inference”, we refer to beranger2021 for a full description.
The methodology is illustrated on the dataset Milan.winter, which consists of air pollution levels recorded in Milan, Italy, over the winter period October 31st–February 27/28th, between December 31st 2001 and December 30th 2017. Here we focus on the and pollutants and use the daily maximum temperature (MaxTemp) as covariate. The data is prepared before estimating extreme quantile regions using fExtDep.np.
R> data(MilanPollution)R> data <- Milan.winter[,c("NO2","SO2", "MaxTemp")]R> data <- data[complete.cases(data),] A quadratic relationship between pollutants and maximum temperature is considered, and the th marginal mean is written as , with , where is the temperature level. The covariate matrix is defined as
R> covar <- cbind(rep(1,nrow(data)), data[,3], data[,3]^2 ) which will be provided to fExtDep.np through the arguments cov1 and cov2. Since a polynomial angular measure in Bernstein form is considered, we specify a prior distribution for the polynomial degree as a negative binomial on with mean and variance and priors for the point masses and respectively as uniform distributions on and , where and are defined as in the previous two sections. These are the default prior distributions in the fExtDep.np routine, therefore the hyper-parameters are specified by
R> hyperparam <- list(mu.nbinom = 6, var.nbinom = 8, a.unif=0, b.unif=0.2) Starting values for the marginal parameters , , the scaling parameter of the sampler on each margin and the polynomial degree are respectively set by the arguments par10, par20, sig10 sig20 and k0. The argument method="Bayesian" indicates that a Bernstein polynomial is used to represent the extremal dependence and that the inference part follows the MCMC scheme detailed in beranger2021. The argument u=TRUE specifies that a censored likelihood approach is applied on raw data with threshold u set to the marginal quantile by default. Recall that by default the marginal distributions are fitted jointly with the dependence (mar.fit=TRUE) but mar.prelim=FALSE indicates that no initial marginal fit is required.
R> pollut <- fExtDep.np(method="Bayesian", data = data[,-3], u=TRUE,+ cov1 = covar, cov2 = covar, mar.prelim=FALSE,+ par10 = c(100,0,0,35,1), par20 = c(20,0,0,20,1),+ sig10 = 0.1, sig20 = 0.1, k0 = 5,+ hyperparam = hyperparam, nsim = 15e+3)
U set to 90% quantile by default on both marginsEstimation of the extremal dependence and margins |====================================================================| 100% The output of the estimation procedure can be visualised and summarised using plot.ExtDep.np and summary.ExtDep by specifying the arguments type="Qsets", out and summary.mcmc. Since covariates are used to model the marginal parameters, we need to provide the covariate levels at which the extreme quantile regions will be computed. In this example, extreme quantile regions are computed at the minimum, median and maximum of the observed daily maximum temperatures. The covariate matrices QatCov1 and QatCov2 should not include an intercept term and contain a maximum of three levels as given by the following code.
R> pollut_sum <- summary_ExtDep(mcmc=pollut, burn=5e+3)R> Temp.seq <- c(min(data[,3]), median(data[,3]), max(data[,3]) )R> QatTemp <- cbind(Temp.seq, Temp.seq^2) We consider representations of quantile regions associated with small probabilities and and specify that we want to include graphical summaries of the extremal dependence as well as displaying the data (dep=TRUE and data). The arguments xlim, ylim and labels are graphical parameters for the production of the quantile regions.
R> pl <- plot.ExtDep.np(out=pollut, type="Qsets", summary.mcmc=pollut_sum,+ QatCov1=QatTemp, P=1/c(600, 1200, 2400), dep=TRUE,+ data=data[,-3], xlim=c(0,800), ylim=c(0,800),+ labels=c(expression(NO[2]), expression(SO[2])))

The output is stored in the object pl which contains a list of dependence quantities in est.out and a list of quantile sets in q.out. In particular, est.out includes the following elements:
-
•
ghat: a matrix with an estimate of the inverse of the angular basic density evaluated at 100 grid points in . Rows gives the posterior -quantile, mean and -quantile.
-
•
Shat_post and Shat: lists where each element is a matrix providing an estimate of the basic set (obtained through ghat). Shat_post considers every posterior samples whereas Shat takes the pointwise -quantile, mean and -quantile.
-
•
nuShat_post and nuShat: two vectors providing the posterior of the basic set measure and its -quantile, mean and -quantile.
The q.out list in pl includes:
-
•
Qset_PA_CovNum_B that is a list of three matrices. Each of them is an estimate of the bivariate extreme quantile . Such regions are computed for the specified probabilities (P) and covariate (QatCov1 and QatCov2). E.g., pl$Qset_P1_CovNum_1 provides an estimate of the () region corresponding to the probability , when the minimum temperature is observed. Each matrix corresponds to the posterior -quantile, mean and -quantile, obtained from Qset_PA_CovNum_B_post.
-
•
Qset_PA_CovNum_B_post that is a (nsim-burn) matrix providing an estimate of the extreme quantile region for each posterior sample.
The argument dep=TRUE produces the top row of Figure 10. The bottom row illustrates the extreme quantile regions corresponding to probabilities (left), (middle) and (right) where the colours indicates different levels of the covariate. Colours of the data (top right), credibility regions and mean from the posterior can be respectively specified by col.data, col.Qfade and col.Qfull (blue to red colour palette by default).
3 Spatial extremes
A convenient tool to statistically model extremes and their dependence over a spatial domain is provided by max-stable processes (e.g. de2006, Ch. 9). Shortly, let , , be independent copies of a stochastic process , with the same finite-dimensional distribution. If there are functions and , for each , such that the finite-dimensional distribution of a limit process , given by {Y(s),s∈D} = lim_n→∞{max_1⩽i⩽n (Xi(s)-bn(s)an(s)), s∈D