F. R. Bach and E. Moulines, Non-Asymptotic Analysis of Stochastic Approximation Algorithms for Machine Learning, NIPS, 2011.
URL : https://hal.archives-ouvertes.fr/hal-00608041

A. Bellet, A. Habrard, and M. Sebban, A Survey on Metric Learning for Feature Vectors and Structured Data, 2013.

A. Bellet, A. Habrard, and M. Sebban, Metric Learning, Synthesis Lectures on Artificial Intelligence and Machine Learning, vol.9, issue.1, 2015.
DOI : 10.2200/S00626ED1V01Y201501AIM030

URL : https://hal.archives-ouvertes.fr/hal-01121733

L. Bottou and O. Bousquet, The Tradeoffs of Large Scale Learning, NIPS, 2007.

S. Clémençon, G. Lugosi, and N. Vayatis, Ranking and empirical risk minimization of U-statistics, Ann. Statist, vol.36, 2008.

S. Clémençon, S. Robbiano, and J. Tressou, Maximal deviations of incomplete U -processes with applications to Empirical Risk Sampling, SDM, 2013.

S. Clémençon, On U-processes and clustering performance, NIPS, pp.37-45, 2011.

S. Clémençon, P. Bertail, and E. Chautru, Scaling up M-estimation via sampling designs: The Horvitz-Thompson stochastic gradient descent, 2014 IEEE International Conference on Big Data (Big Data), 2014.
DOI : 10.1109/BigData.2014.7004208

A. Defazio, F. Bach, and S. Lacoste-julien, SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives, NIPS, 2014.
URL : https://hal.archives-ouvertes.fr/hal-01016843

B. Delyon, Stochastic Approximation with Decreasing Gain: Convergence and Asymptotic Theory, 2000.

G. Fort, Central limit theorems for stochastic approximation with controlled Markov chain dynamics, ESAIM: Probability and Statistics, vol.19, 2014.
DOI : 10.1051/ps/2014013

URL : https://hal.archives-ouvertes.fr/hal-00861097

J. Fürnkranz, E. Hüllermeier, and S. Vanderlooy, Binary Decomposition Methods for Multipartite Ranking, ECML/PKDD, pp.359-374, 2009.
DOI : 10.1016/j.patrec.2007.07.019

A. Herschtal and B. Raskutti, Optimising area under the ROC curve using gradient descent, Twenty-first international conference on Machine learning , ICML '04, p.49, 2004.
DOI : 10.1145/1015330.1015366

S. Janson, The asymptotic distributions of incomplete U-statistics, Zeitschrift f???r Wahrscheinlichkeitstheorie und Verwandte Gebiete, vol.49, issue.4, pp.495-505, 1984.
DOI : 10.1007/BF00531887

R. Johnson and T. Zhang, Accelerating Stochastic Gradient Descent using Predictive Variance Reduction, NIPS, pp.315-323, 2013.

P. Kar, B. Sriperumbudur, P. Jain, and H. Karnick, On the Generalization Ability of Online Learning Algorithms for Pairwise Loss Functions, ICML, 2013.

H. J. Kushner and G. Yin, Stochastic approximation and recursive algorithms and applications, 2003.

N. , L. Roux, M. W. Schmidt, and F. Bach, A Stochastic Gradient Method with an Exponential Convergence Rate for Finite Training Sets, NIPS, 2012.
URL : https://hal.archives-ouvertes.fr/hal-00674995

A. J. Lee, U-Statistics: Theory and Practice, 1990.

D. Needell, R. Ward, and N. Srebro, Stochastic gradient descent, weighted sampling, and the randomized Kaczmarz algorithm, NIPS, pp.1017-1025, 2014.
DOI : 10.1007/s10107-015-0864-7

A. Nemirovski, A. Juditsky, G. Lan, and A. Shapiro, Robust Stochastic Approximation Approach to Stochastic Programming, SIAM Journal on Optimization, vol.19, issue.4, pp.1574-1609, 2009.
DOI : 10.1137/070704277

URL : https://hal.archives-ouvertes.fr/hal-00976649

Y. Nesterov, Introductory lectures on convex optimization, 2004.
DOI : 10.1007/978-1-4419-8853-9

M. Norouzi, D. J. Fleet, and R. Salakhutdinov, Hamming Distance Metric Learning, NIPS, pp.1070-1078, 2012.

M. Pelletier, Weak convergence rates for stochastic approximation with application to multiple targets and simulated annealing, The Annals of Applied Probability, vol.8, issue.1, 1998.
DOI : 10.1214/aoap/1027961032

Q. Qian, R. Jin, J. Yi, L. Zhang, and S. Zhu, Efficient distance metric learning by adaptive sampling and mini-batch stochastic gradient descent (SGD), Machine Learning, vol.4, issue.1, pp.353-372, 2015.
DOI : 10.1007/s10994-014-5456-x

P. Zhao and T. Zhang, Stochastic Optimization with Importance Sampling for Regularized Loss Minimization, ICML, 2015.