A. Agarwal and L. Bottou, A lower bound for the optimization of finite sums, Proc. International Conference on Machine Learning (ICML), 2015.

F. Bach, R. Jenatton, J. Mairal, and G. Obozinski, Optimization with Sparsity-Inducing Penalties, Machine Learning, pp.1-106, 2012.
DOI : 10.1561/2200000015

URL : https://hal.archives-ouvertes.fr/hal-00613125

H. H. Bauschke and P. L. Combettes, Convex Analysis and Monotone Operator Theory in Hilbert Spaces, 2011.
URL : https://hal.archives-ouvertes.fr/hal-00643354

A. Beck and M. Teboulle, A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems, SIAM Journal on Imaging Sciences, vol.2, issue.1, pp.183-202, 2009.
DOI : 10.1137/080716542

D. P. Bertsekas, Convex Optimization Algorithms, Athena Scientific, 2015.

A. J. Defazio, F. Bach, and S. Lacoste-julien, SAGA: A fast incremental gradient method with support for non-strongly convex composite objectives, Adv. Neural Information Processing Systems (NIPS), 2014.
URL : https://hal.archives-ouvertes.fr/hal-01016843

A. J. Defazio, T. S. Caetano, and J. Domke, Finito: A faster, permutable incremental gradient method for big data problems, Proc. International Conference on Machine Learning (ICML), 2014.

R. Frostig, R. Ge, S. M. Kakade, and A. Sidford, Un-regularizing: approximate proximal point algorithms for empirical risk minimization, Proc. International Conference on Machine Learning (ICML), 2015.

O. Güler, New Proximal Point Algorithms for Convex Minimization, SIAM Journal on Optimization, vol.2, issue.4, pp.649-664, 1992.
DOI : 10.1137/0802032

B. He and X. Yuan, An Accelerated Inexact Proximal Point Algorithm for Convex Minimization, Journal of Optimization Theory and Applications, vol.2, issue.2, pp.536-548, 2012.
DOI : 10.1007/s10957-011-9948-6

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.414.2193

J. Hiriart-urruty and C. Lemaréchal, Convex Analysis and Minimization Algorithms I, 1996.
DOI : 10.1007/978-3-662-02796-7

A. Juditsky and A. Nemirovski, First order methods for nonsmooth convex large-scale optimization. Optimization for Machine Learning, 2012.
URL : https://hal.archives-ouvertes.fr/hal-00981863

G. Lan, An optimal randomized incremental gradient method, 2015.

A. Nemirovski, A. Juditsky, G. Lan, and A. Shapiro, Robust Stochastic Approximation Approach to Stochastic Programming, SIAM Journal on Optimization, vol.19, issue.4, pp.1574-1609, 2009.
DOI : 10.1137/070704277

URL : https://hal.archives-ouvertes.fr/hal-00976649

Y. Nesterov, A method of solving a convex programming problem with convergence rate, Soviet Mathematics Doklady, vol.27, issue.1 22, pp.372-376, 1983.

Y. Nesterov, Introductory Lectures on Convex Optimization: A Basic Course, 2004.
DOI : 10.1007/978-1-4419-8853-9

Y. Nesterov, Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems, SIAM Journal on Optimization, vol.22, issue.2, pp.341-362, 2012.
DOI : 10.1137/100802001

Y. Nesterov, Gradient methods for minimizing composite functions, Mathematical Programming, pp.125-161, 2013.
DOI : 10.1007/s10107-012-0629-5

N. Parikh and S. P. Boyd, Proximal Algorithms, Foundations and Trends?? in Optimization, vol.1, issue.3, pp.123-231, 2014.
DOI : 10.1561/2400000003

P. Richtárik and M. Taká?, Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function, Mathematical Programming, pp.1-38, 2014.
DOI : 10.1007/s10107-012-0614-z

S. Salzo and S. Villa, Inexact and accelerated proximal point algorithms, Journal of Convex Analysis, vol.19, issue.4, pp.1167-1192, 2012.

M. Schmidt, N. L. Roux, and F. Bach, Convergence rates of inexact proximal-gradient methods for convex optimization, Adv. Neural Information Processing Systems (NIPS), 2011.
URL : https://hal.archives-ouvertes.fr/inria-00618152

M. Schmidt, N. L. Roux, and F. Bach, Minimizing finite sums with the stochastic average gradient, 2013.
URL : https://hal.archives-ouvertes.fr/hal-00860051

S. Shalev-shwartz and T. Zhang, Proximal stochastic dual coordinate ascent, 2012.

S. Shalev-shwartz and T. Zhang, Accelerated proximal stochastic dual coordinate ascent for regularized loss minimization, Mathematical Programming, 2015.
DOI : 10.1007/s10107-014-0839-0

L. Xiao and T. Zhang, A Proximal Stochastic Gradient Method with Progressive Variance Reduction, SIAM Journal on Optimization, vol.24, issue.4, pp.2057-2075, 2014.
DOI : 10.1137/140961791

Y. Zhang and L. Xiao, Stochastic primal-dual coordinate method for regularized empirical risk minimization, Proc. International Conference on Machine Learning (ICML), 2015.