Principal Component Analysis for Multivariate Extremes

Abstract : The first order behavior of multivariate heavy-tailed random vectors above large radial thresholds is ruled by a limit measure in a regular variation framework. For a high dimensional vector, a reasonable assumption is that the support of this measure is concentrated on a lower dimensional subspace, meaning that certain linear combinations of the components are much likelier to be large than others. Identifying this subspace and thus reducing the dimension will facilitate a refined statistical analysis. In this work we apply Principal Component Analysis (PCA) to a re-scaled version of radially thresholded observations. Within the statistical learning framework of empirical risk minimization, our main focus is to analyze the squared reconstruction error for the exceedances over large radial thresholds. We prove that the empirical risk converges to the true risk, uniformly over all projection subspaces. As a consequence, the best projection subspace is shown to converge in probability to the optimal one, in terms of the Hausdorff distance between their intersections with the unit sphere. In addition, if the exceedances are re-scaled to the unit ball, we obtain finite sample uniform guarantees to the reconstruction error pertaining to the estimated projection sub-space. Numerical experiments illustrate the relevance of the proposed framework for practical purposes.
Complete list of metadatas

https://hal.archives-ouvertes.fr/hal-02164305
Contributor : Anne Sabourin <>
Submitted on : Tuesday, June 25, 2019 - 12:13:08 AM
Last modification on : Monday, July 8, 2019 - 2:59:19 PM

Files

PCA_final_arxiv.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-02164305, version 1
  • ARXIV : 1906.11043

Citation

Holger Drees, Anne Sabourin. Principal Component Analysis for Multivariate Extremes. 2019. ⟨hal-02164305⟩

Share

Metrics

Record views

17

Files downloads

20