G. Biau and B. Cadre, Optimization by gradient boosting, 2017.
URL : https://hal.archives-ouvertes.fr/hal-01562618

G. Biau, A. Fischer, B. Guedj, and J. D. Malley, COBRA: A combined regression strategy, Journal of Multivariate Analysis, vol.146, pp.18-28, 2016.
URL : https://hal.archives-ouvertes.fr/hal-01361789

G. Biau, B. Cadre, and L. Rouvière, Accelerated Gradient Boosting, 2018.

L. Breiman, Arcing the Edge, Statistics Department, 1997.

L. Breiman, Arcing classifier (with discussion and a rejoinder by the author), The Annals of Statistics, vol.26, issue.3, pp.801-849, 1998.

L. Breiman, Prediction Games and Arcing Algorithms, Neural Computation, vol.11, issue.7, pp.1493-1517, 1999.

L. Breiman, Some Infinite Theory for Predictor Ensembles, Statistics Department, 2000.

L. Breiman, Random Forests, Machine Learning, vol.45, pp.5-32, 2001.

L. Breiman, Population theory for boosting ensembles, The Annals of Statistics, vol.32, issue.1, pp.1-11, 2004.

P. Bühlmann and T. Hothorn, Boosting Algorithms: Regularization, Prediction and Model Fitting, Statistical Science, vol.22, issue.4, pp.477-505, 2007.

T. Chen and C. Guestrin, XGBoost: A Scalable Tree Boosting System, Proceedings of the 22Nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp.785-794, 2016.

P. Combettes and V. Wajs, Signal Recovery by Proximal Forward-Backward Splitting, Multiscale Modeling & Simulation, vol.4, issue.4, pp.1168-1200, 2005.
URL : https://hal.archives-ouvertes.fr/hal-00017649

Y. Freund, Boosting a Weak Learning Algorithm by Majority, Information and Computation, vol.121, issue.2, pp.256-285, 1995.

Y. Freund and R. E. Schapire, Experiments with a New Boosting Algorithm, Proceedings of the Thirteenth International Conference on International Conference on Machine Learning, 1996.

Y. Freund and R. E. Schapire, A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting, Journal of Computer and System Sciences, vol.55, issue.1, pp.119-139, 1997.

J. Friedman, Greedy function approximation: A gradient boosting machine, The Annals of Statistics, vol.29, issue.5, pp.1189-1232, 2001.

J. Friedman, Stochastic gradient boosting, Computational Statistics & Data Analysis, vol.38, issue.4, pp.367-378, 2002.

J. Friedman, T. Hastie, and R. Tibshirani, Additive logistic regression: a statistical view of boosting (with discussion and a rejoinder by the authors), The Annals of Statistics, vol.28, issue.2, pp.337-407, 2000.

A. Grubb and J. A. Bagnell, Generalized Boosting Algorithms for Convex Optimization, Proceedings of The 28th International Conference on Machine Learning, 2011.

M. Jaggi, Revisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization, Proceedings of The 30th International Conference on Machine Learning, pp.427-435, 2013.

J. Lin, L. Rosasco, and D. Zhou, Iterative Regularization for Learning with Convex Loss Functions, Journal of Machine Learning Research, vol.17, issue.77, pp.1-38, 2016.

L. Mason, J. Baxter, P. L. Bartlett, and M. Frean, Boosting Algorithms as Gradient Descent, Advances in Neural Information Processing Systems, pp.512-518, 2000.

L. Mason, J. Baxter, P. L. Bartlett, and M. Frean, Functional gradient techniques for combining hypotheses, Advances in Large Margin Classifiers, pp.221-246, 2000.

R. Meir and G. Rätsch, An Introduction to Boosting and Leveraging, Advanced Lectures on Machine Learning, pp.118-183, 2003.

Y. Nesterov, A method of solving a convex programming problem with convergence rate 0(1/k 2 ), Soviet Mathematics Doklady, p.27, 1983.

Y. Nesterov, Introductory Lectures on Convex Optimization: A Basic Course, 2004.

P. Ochs, Y. Chen, T. Brox, and T. Pock, iPiano: Inertial Proximal Algorithm for Nonconvex Optimization, SIAM Journal on Imaging Sciences, 2014.

F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion et al., Scikit-learn: Machine Learning in Python, Journal of Machine Learning Research, vol.12, pp.2825-2830, 2011.
URL : https://hal.archives-ouvertes.fr/hal-00650905

. Tyrrell-rockafellar, Monotone operators and the proximal point algorithm, SIAM journal on control and optimization, vol.14, issue.5, pp.877-898, 1976.

G. Rätsch, S. Mika, and M. K. Warmuth, On the Convergence of Leveraging, Advances in Neural Information Processing Systems, pp.487-494, 2002.

R. E. Schapire, The strength of weak learnability, Machine Learning, vol.5, pp.197-227, 1990.

V. N. Temlyakov, Greedy expansions in convex optimization, Proceedings of the Steklov Institute of Mathematics, 2012.

C. Wang, Y. Wang, W. E. , and R. Schapire, Functional Frank-Wolfe Boosting for General Loss Functions, 2015.

T. Zhang, A General Greedy Approximation Algorithm with Applications, Advances in Neural Information Processing Systems, vol.14, pp.1065-1072, 2002.

T. Zhang, Sequential greedy approximation for certain convex optimization problems, IEEE Transactions on Information Theory, vol.49, issue.3, pp.682-691, 2003.