Sequential dimension reduction for learning features of expensive black-box functions - Archive ouverte HAL Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2018

Sequential dimension reduction for learning features of expensive black-box functions

Résumé

Learning a feature of an expensive black-box function (optimum, contour line,...) is a difficult task when the dimension increases. A classical approach is two-stage. First, sensitivity analysis is performed to reduce the dimension of the input variables. Second, the feature is estimated by considering only the selected influential variables. This approach can be computationally expensive and may lack flexibility since dimension reduction is done once and for all. In this paper, we propose a so called Split-and-Doubt algorithm that performs sequentially both dimension reduction and feature oriented sampling. The 'split' step identifies influential variables. This selection relies on new theoretical results on Gaussian process regression. We prove that large correlation lengths of covariance functions correspond to inactive variables. Then, in the 'doubt' step, a doubt function is used to update the subset of influential variables. Numerical tests show the efficiency of the Split-and-Doubt algorithm.
Fichier principal
Vignette du fichier
split_and_doubt (3).pdf (1.37 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01688329 , version 1 (19-01-2018)
hal-01688329 , version 2 (27-02-2019)

Identifiants

  • HAL Id : hal-01688329 , version 1

Citer

Malek Ben Salem, François Bachoc, Olivier Roustant, Fabrice Gamboa, Lionel Tomaso. Sequential dimension reduction for learning features of expensive black-box functions. 2018. ⟨hal-01688329v1⟩
574 Consultations
1517 Téléchargements

Partager

Gmail Facebook X LinkedIn More