On the approximation of extreme quantiles with ReLU neural networks.

Authors
Publication date
2021
Publication type
Proceedings Article
Summary Feedforward neural networks based on rectified linear units (ReLU) cannot efficiently approximate quantile functions which are not bounded in the Fréchet maximum domain of attraction. We thus propose a new parametrization for the generator of a generative adversarial network (GAN) adapted to this framework of heavy-tailed distributions. We provide an analysis of the uniform error between the extreme quantile and its GAN approximation. It appears that the rate of convergence of the error is mainly driven by the second-order parameter of the data distribution. The above results are illustrated on simulated data and real financial data.
Topics of the publication
  • ...
  • No themes identified
Themes detected by scanR from retrieved publications. For more information, see https://scanr.enseignementsup-recherche.gouv.fr