Sequential dimension reduction for learning features of expensive black-box functions

Abstract : Learning a feature of an expensive black-box function (optimum, contour line,...) is a difficult task when the dimension increases. A classical approach is two-stage. First, sensitivity analysis is performed to reduce the dimension of the input variables. Second, the feature is estimated by considering only the selected influential variables. This approach can be computationally expensive and may lack flexibility since dimension reduction is done once and for all. In this paper, we propose a so called Split-and-Doubt algorithm that performs sequentially both dimension reduction and feature oriented sampling. The 'split' step identifies influential variables. This selection relies on new theoretical results on Gaussian process regression. We prove that large correlation lengths of covariance functions correspond to inactive variables. Then, in the 'doubt' step, a doubt function is used to update the subset of influential variables. Numerical tests show the efficiency of the Split-and-Doubt algorithm.
Type de document :
Pré-publication, Document de travail
2019
Liste complète des métadonnées


https://hal.archives-ouvertes.fr/hal-01688329
Contributeur : Malek Ben Salem <>
Soumis le : mercredi 27 février 2019 - 11:09:48
Dernière modification le : vendredi 15 mars 2019 - 01:14:46

Identifiants

  • HAL Id : hal-01688329, version 2

Citation

Malek Ben Salem, François Bachoc, Olivier Roustant, Fabrice Gamboa, Lionel Tomaso. Sequential dimension reduction for learning features of expensive black-box functions. 2019. 〈hal-01688329v2〉

Partager

Métriques

Consultations de la notice

38

Téléchargements de fichiers

100