On the approximation of extreme quantiles with ReLU neural networks - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2021

On the approximation of extreme quantiles with ReLU neural networks

Résumé

Feedforward neural networks based on rectified linear units (ReLU) cannot efficiently approximate quantile functions which are not bounded in the Fréchet maximum domain of attraction. We thus propose a new parametrization for the generator of a generative adversarial network (GAN) adapted to this framework of heavy-tailed distributions. We provide an analysis of the uniform error between the extreme quantile and its GAN approximation. It appears that the rate of convergence of the error is mainly driven by the second-order parameter of the data distribution. The above results are illustrated on simulated data and real financial data.
Fichier non déposé

Dates et versions

hal-03301431 , version 1 (27-07-2021)

Identifiants

  • HAL Id : hal-03301431 , version 1

Citer

Michaël Allouche, Stéphane Girard, Emmanuel Gobet. On the approximation of extreme quantiles with ReLU neural networks. EVA 2021 - 12th International Conference on Extreme Value Analysis, Jun 2021, Edinburgh / Virtual, United Kingdom. ⟨hal-03301431⟩
109 Consultations
0 Téléchargements

Partager

Gmail Facebook X LinkedIn More