D. Calandriello, A. Lazaric, and M. Valko, Second-order kernel online convex optimization with adaptive sketching, International Conference on Machine Learning, 2017.
URL : https://hal.archives-ouvertes.fr/hal-01537799

G. Cavallanti, N. Cesa-bianchi, and C. Gentile, Tracking the best hyperplane with a simple budget Perceptron, Machine Learning, pp.143-167, 2007.
DOI : 10.1007/s10994-007-5003-0

B. Michael, C. Cohen, J. Musco, and . Pachocki, Online row sampling. International Workshop on Approximation, Randomization, and Combinatorial Optimization APPROX, 2016.

O. Dekel, S. Shalev-shwartz, and Y. Singer, The Forgetron: A Kernel-Based Perceptron on a Budget, SIAM Journal on Computing, vol.37, issue.5, pp.1342-1372, 2008.
DOI : 10.1137/060666998

M. Ghashami, E. Liberty, M. Jeff, . Phillips, P. David et al., Frequent Directions: Simple and Deterministic Matrix Sketching, SIAM Journal on Computing, vol.45, issue.5, pp.1762-1792, 2016.
DOI : 10.1137/15M1009718

E. Hazan, A. Kalai, S. Kale, and A. Agarwal, Logarithmic regret algorithms for online convex optimization, Conference on Learning Theory, 2006.
DOI : 10.1007/11776420_37

URL : http://www.cs.princeton.edu/~satyen/papers/HKKA2006.pdf

W. He and J. T. Kwok, Simple randomized algorithms for online learning with kernels, Neural Networks, vol.60, pp.17-24, 2014.
DOI : 10.1016/j.neunet.2014.07.006

URL : http://www.cse.ust.hk/~jamesk/papers/nn14.pdf

J. Kivinen, A. J. Smola, and R. C. Williamson, Online Learning with Kernels, IEEE Transactions on Signal Processing, vol.52, issue.8, p.52, 2004.
DOI : 10.1109/TSP.2004.830991

Q. Le, T. Sarlós, and A. J. Smola, Fastfood -Approximating kernel expansions in loglinear time, International Conference on Machine Learning, 2013.

T. Le, T. Nguyen, V. Nguyen, and D. Phung, Dual Space Gradient Descent for Online Learning, Neural Information Processing Systems, 2016.

J. Lu, C. H. Steven, J. Hoi, P. Wang, Z. Zhao et al., Large scale online kernel learning, Journal of Machine Learning Research, vol.17, issue.47, pp.1-43, 2016.

H. Luo, A. Agarwal, N. Cesa-bianchi, and J. Langford, Efficient second-order online learning via sketching, Neural Information Processing Systems, 2016.

F. Orabona, J. Keshet, and B. Caputo, The projectron, Proceedings of the 25th international conference on Machine learning, ICML '08, 2008.
DOI : 10.1145/1390156.1390247

Y. Sun, J. Schmidhuber, and F. J. Gomez, On the size of the online kernel sparsification dictionary, International Conference on Machine Learning, 2012.

Z. Wang, K. Crammer, and S. Vucetic, Breaking the curse of kernelization: Budgeted stochastic gradient descent for large-scale svm training, Journal of Machine Learning Research, vol.13, pp.3103-3131, 2012.

J. Andrew, S. Wathen, and . Zhu, On spectral distribution of kernel matrices related to radial basis functions, Numerical Algorithms, vol.70, issue.4, pp.709-726, 2015.

C. Williams and M. Seeger, Using the Nyström method to speed up kernel machines, Neural Information Processing Systems, 2001.

Y. Xu, H. Yang, L. Zhang, and T. Yang, Efficient non-oblivious randomized reduction for risk minimization with improved excess risk guarantee, AAAI Conference on Artificial Intelligence, 2017.

T. Yang, Y. Li, M. Mahdavi, R. Jin, and Z. Zhou, Nyström method vs random fourier features: A theoretical and empirical comparison, Neural Information Processing Systems, 2012.

Y. Yang, M. Pilanci, and M. J. Wainwright, Randomized sketches for kernels: Fast and optimal nonparametric regression, The Annals of Statistics, vol.45, issue.3, 2017.
DOI : 10.1214/16-AOS1472

URL : http://arxiv.org/abs/1501.06195

P. Zhao, J. Wang, P. Wu, R. Jin, C. Steven et al., Fast bounded online gradient descent algorithms for scalable kernel-based online learning, International Conference on Machine Learning, 2012.

F. Zhdanov and Y. Kalnishkan, An identity for kernel ridge regression, Algorithmic Learning Theory, 2010.
DOI : 10.1007/978-3-642-16108-7_32

URL : http://arxiv.org/pdf/1112.1390.pdf

C. Zhu and H. Xu, Online gradient descent in function space, 2015.

M. Zinkevich, Online convex programming and generalized infinitesimal gradient ascent, International Conference on Machine Learning, 2003.