C. Cortes and V. Vapnik, Support-vector networks, Machine Learning, vol.1, issue.3, pp.273-2971022627411411, 1995.
DOI : 10.1007/BF00994018

O. Chapelle, V. Vapnik, O. Bousquet, and S. Mukherjee, Choosing multiple parameters for support vector machines Available: https://doi.org/10, Machine Learning, pp.131-1591012450327387, 1023.

C. M. Bishop, Pattern Recognition and Machine Learning (Information Science and Statistics), 2006.

Y. Lecun, L. Bottou, G. Orr, and K. Mller, Efficient Backprop, ser, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics, vol.7700, pp.9-48, 2012.

G. E. Hinton, A Practical Guide to Training Restricted Boltzmann Machines, Neural Networks: Tricks of the Trade, pp.599-619, 2012.
DOI : 10.1145/1390156.1390290

URL : http://learning.cs.toronto.edu/%7Ehinton/absps/guideTR.pdf

, Dynamic Early Stopping Criterion for Random Search in SVM Hyperparameter Optimization 11

S. Smusz, W. M. Czarnecki, D. Warszycki, and A. J. Bojarski, Exploiting uncertainty measures in compounds activity prediction using support vector machines, Bioorganic & Medicinal Chemistry Letters, vol.25, issue.1, pp.100-105, 2015.
DOI : 10.1016/j.bmcl.2014.11.005

J. A. Nelder and R. Mead, A Simplex Method for Function Minimization, The Computer Journal, vol.7, issue.4, pp.308-313, 1965.
DOI : 10.1093/comjnl/7.4.308

S. Kirkpatrick, Optimization by simulated annealing: Quantitative studies, Journal of Statistical Physics, vol.21, issue.5-6, pp.975-986, 1984.
DOI : 10.1007/BF01009452

N. Hansen, S. D. Muller, and P. Koumoutsakos, Reducing the Time Complexity of the Derandomized Evolution Strategy with Covariance Matrix Adaptation (CMA-ES), Evolutionary Computation, vol.11, issue.1, pp.1-18, 2003.
DOI : 10.1162/106365601750190398

C. Thornton, F. Hutter, H. H. Hoos, and K. Leyton-brown, Auto-WEKA, Proceedings of the 19th ACM SIGKDD international conference on Knowledge discovery and data mining, KDD '13, pp.847-855, 2013.
DOI : 10.1145/2487575.2487629

J. Bergstra, R. Bardenet, Y. Bengio, B. Kgl, J. Shawe-taylor et al., Algorithms for hyper-parameter optimization, NIPS, pp.2546-2554, 2011.
URL : https://hal.archives-ouvertes.fr/hal-00642998

J. Bergstra and Y. Bengio, Random Search for Hyper-Parameter Optimization, Journal of Machine Learning Research, vol.13, pp.281-305, 2012.

J. Lemley, F. Jagodzinski, and R. Andonie, Big Holes in Big Data: A Monte Carlo Algorithm for Detecting Large Hyper-Rectangles in High Dimensional Data, 2016 IEEE 40th Annual Computer Software and Applications Conference (COMPSAC), pp.563-571, 2016.
DOI : 10.1109/COMPSAC.2016.73

C. Chang and C. Lin, LIBSVM, ACM Transactions on Intelligent Systems and Technology, vol.2, issue.3, pp.1-27, 2011.
DOI : 10.1145/1961189.1961199

F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion et al., Scikit-learn: Machine learning in Python, Journal of Machine Learning Research, vol.12, pp.2825-2830, 2011.
URL : https://hal.archives-ouvertes.fr/hal-00650905

J. Snoek, H. Larochelle, R. P. Adams, F. Pereira, C. J. Burges et al., Practical bayesian optimization of machine learning algorithms Available: http://papers.nips.cc/paper/4522-practical-bayesian-optimization-of-machine- learning-algorithms.pdf 17. R. Martinez-Cantin Bayesopt: A bayesian optimization library for nonlinear optimization, experimental design and bandits, Advances in Neural Information Processing Systems 25, pp.2951-29597430, 1405.

L. Kotthoff, C. Thornton, H. H. Hoos, F. Hutter, and K. Leyton-brown, Auto- WEKA 2.0: Automatic model selection and hyperparameter optimization in WEKA, Journal of Machine Learning Research, vol.18, issue.25, pp.1-516, 2017.

A. C?t?lin, F. , and R. Andonie,

M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann et al., The WEKA data mining software, ACM SIGKDD Explorations Newsletter, vol.11, issue.1, pp.10-18, 2009.
DOI : 10.1145/1656274.1656278

J. Bergstra, B. Komer, C. Eliasmith, D. Yamins, and D. D. Cox, Hyperopt: a Python library for model selection and hyperparameter optimization, Computational Science & Discovery, vol.8, issue.1, pp.140081749-4699, 2015.
DOI : 10.1088/1749-4699/8/1/014008

URL : http://iopscience.iop.org/article/10.1088/1749-4699/8/1/014008/pdf

M. Claesen, J. Simm, D. Popovic, Y. Moreau, and B. D. Moor, Easy Hyperparameter Search Using Optunity

. Google, The Go programming language Available: https://golang.org, 2007.

M. J. Quinn, Parallel Programming in C with MPI and OpenMP, 2003.

M. Sokolova and G. Lapalme, A systematic analysis of performance measures for classification tasks, Information Processing & Management, vol.45, issue.4, pp.427-437, 2009.
DOI : 10.1016/j.ipm.2009.03.002

J. Demsar, Statistical comparisons of classifiers over multiple data sets, Journal of Machine Learning Research, vol.7, pp.1-30, 2006.

M. Friedman, A Comparison of Alternative Tests of Significance for the Problem of $m$ Rankings, The Annals of Mathematical Statistics, vol.11, issue.1, pp.86-92, 1940.
DOI : 10.1214/aoms/1177731944

R. Iman and J. Davenport, Approximations of the critical region of the fbietkan statistic, Communications in Statistics - Theory and Methods, vol.2, issue.6, pp.571-595, 1980.
DOI : 10.2307/3002019

P. Nemenyi, Distribution-free Multiple Comparisons Available: https, 1963.

S. Holm, A simple sequentially rejective multiple test procedure, Scandinavian Journal of Statistics, vol.6, pp.65-70, 1979.