Information-theoretic lower bounds on the oracle complexity of convex optimization, IEEE Transactions on Information Theory, vol.58, issue.5, pp.3235-3249, 2012. ,
Katyusha: The first direct acceleration of stochastic gradient methods, Proceedings of Symposium on Theory of Computing (STOC, 2017. ,
Dimension-free iteration complexity of finite sum optimization problems, Advances in Neural Information Processing Systems (NIPS), 2016. ,
Estimate sequence methods: extensions and approximations. ETH technical report, 2009. ,
A fast iterative shrinkage-thresholding algorithm for linear inverse problems, SIAM Journal on Imaging Sciences, vol.2, issue.1, pp.183-202, 2009. ,
Stochastic optimization with variance reduction for infinite datasets with finite-sum structure, Advances in Neural Information Processing Systems (NIPS), 2017. ,
URL : https://hal.archives-ouvertes.fr/hal-01375816
Optimization methods for large-scale machine learning, SIAM Review, vol.60, issue.2, pp.223-311, 2018. ,
Stochastic dual coordinate ascent with adaptive probabilities, International Conference on Machine Learning (ICML), 2015. ,
Saga: A fast incremental gradient method with support for nonstrongly convex composite objectives, Advances in Neural Information Processing Systems (NIPS), 2014. ,
URL : https://hal.archives-ouvertes.fr/hal-01016843
Finito: A faster, permutable incremental gradient method for big data problems, Proceedings of the International Conferences on Machine Learning (ICML), 2014. ,
Optimal stochastic approximation algorithms for strongly convex stochastic composite optimization I: A generic algorithmic framework, SIAM Journal on Optimization, vol.22, issue.4, pp.1469-1492, 2012. ,
Optimal stochastic approximation algorithms for strongly convex stochastic composite optimization II: Shrinking procedures and optimal algorithms, SIAM Journal on Optimization, vol.23, issue.4, pp.2061-2089, 2013. ,
Convex analysis and minimization algorithms. II, 1996. ,
Variance reduced stochastic gradient descent with neighbors, Advances in Neural Information Processing Systems (NIPS), 2015. ,
URL : https://hal.archives-ouvertes.fr/hal-01248672
Accelerated gradient methods for stochastic optimization and online learning, Advances in Neural Information Processing Systems (NIPS), 2009. ,
Accelerating stochastic gradient descent using predictive variance reduction, Advances in Neural Information Processing Systems (NIPS), 2013. ,
An optimal method for stochastic composite optimization, Mathematical Programming, vol.133, issue.1, pp.365-397, 2012. ,
An optimal randomized incremental gradient method, Mathematical Programming, vol.171, issue.1-2, pp.167-215, 2018. ,
Catalyst acceleration for first-order convex optimization: from theory to practice, Journal of Machine Learning Research (JMLR), vol.18, issue.212, pp.1-54, 2018. ,
URL : https://hal.archives-ouvertes.fr/hal-01664934
Incremental majorization-minimization optimization with application to large-scale machine learning, SIAM Journal on Optimization, vol.25, issue.2, pp.829-855, 2015. ,
End-to-end kernel learning with supervised convolutional kernel networks, Advances in Neural Information Processing Systems (NIPS), 2016. ,
Sparse modeling for image and vision processing. Foundations and Trends in Computer Graphics and Vision, vol.8, pp.85-283, 2014. ,
URL : https://hal.archives-ouvertes.fr/hal-01081139
Stability selection, Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol.72, issue.4, pp.417-473, 2010. ,
Fonctions convexes duales et points proximaux dans un espace hilbertien, CR Acad. Sci. Paris Sér. A Math, vol.255, pp.2897-2899, 1962. ,
URL : https://hal.archives-ouvertes.fr/hal-01867195
Proximité et dualité dans un espace hilbertien, Bull. Soc. Math. France, vol.93, issue.2, pp.273-299, 1965. ,
Robust stochastic approximation approach to stochastic programming, SIAM Journal on Optimization, vol.19, issue.4, pp.1574-1609, 2009. ,
URL : https://hal.archives-ouvertes.fr/hal-00976649
A method of solving a convex programming problem with convergence rate O(1/k 2 ), Soviet Mathematics Doklady, vol.27, issue.2, pp.372-376, 1983. ,
Introductory Lectures on Convex Optimization: A Basic Course, 2004. ,
Gradient methods for minimizing composite functions, Mathematical Programming, vol.140, issue.1, pp.125-161, 2013. ,
Cubic regularization of Newton method and its global performance, Mathematical Programming, vol.108, issue.1, pp.177-205, 2006. ,
Stochastic proximal gradient descent with acceleration techniques, Advances in Neural Information Processing Systems (NIPS), 2014. ,
Non-uniform stochastic average gradient method for training conditional random fields, Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS), 2015. ,
Minimizing finite sums with the stochastic average gradient, Mathematical Programming, vol.162, issue.1-2, pp.83-112, 2017. ,
URL : https://hal.archives-ouvertes.fr/hal-00860051
Accelerated proximal stochastic dual coordinate ascent for regularized loss minimization, Mathematical Programming, vol.155, issue.1, pp.105-145, 2016. ,
Dropout: a simple way to prevent neural networks from overfitting, Journal of Machine Learning Research, vol.15, issue.1, pp.1929-1958, 2014. ,
The nature of statistical learning theory, 2000. ,
Privacy aware learning, Advances in Neural Information Processing Systems (NIPS), 2012. ,
A proximal stochastic gradient method with progressive variance reduction, SIAM Journal on Optimization, vol.24, issue.4, pp.2057-2075, 2014. ,