J. L. Ba, J. R. Kiros, and G. Hinton, , 2016.

A. Bibi, M. Alfadly, and B. Ghanem, Analytic Expressions for Probabilistic Moments of PL-DNN with Gaussian Input, The IEEE Conference on Computer Vision and Pattern Recognition, 2018.

Y. Cho and L. K. Saul, Kernel methods for deep learning, Advances in Neural Information Processing Systems, pp.342-350, 2009.

A. Damianou and N. Lawrence, Deep Gaussian processes, Artificial Intelligence and Statistics, pp.207-215, 2013.

D. Duvenaud, O. Rippel, R. Adams, and Z. Ghahramani, Avoiding pathologies in very deep networks, Artificial Intelligence and Statistics, pp.202-210, 2014.

R. M. Neal, Bayesian learning for neural networks, vol.118, 1996.

N. G. Polson and V. Sokolov, Deep learning: A Bayesian perspective, Bayesian Analysis, vol.12, issue.4, pp.1275-1304, 2017.

C. E. Rasmussen and C. K. Williams, Gaussian Processes for Machine Learning, 2006.

H. Rinne, The Weibull distribution: a handbook, 2008.

Y. Saatci, A. G. Wilson, and G. Bayesian, Advances in Neural Information Processing Systems, pp.3622-3631, 2017.

D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre et al., Mastering the game of go with deep neural networks and tree search, Nature, vol.529, issue.7587, pp.484-489, 2016.

M. Sklar, Fonctions de repartition an dimensions et leurs marges, Publ. inst. statist. univ. Paris, vol.8, pp.229-231, 1959.

N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, Dropout: A simple way to prevent neural networks from overfitting, The Journal of Machine Learning Research, vol.15, issue.1, pp.1929-1958, 2014.

R. Tibshirani, Regression shrinkage and selection via the Lasso, Journal of the Royal Statistical Society. Series B (Methodological), pp.267-288, 1996.