H. Bakircio?-glu and T. Koçak, Survey of random neural network applications, European Journal of Operational Research, vol.126, issue.2, pp.319-330, 2000.
DOI : 10.1016/S0377-2217(99)00481-6

S. Basterrech, C. Fyfe, and G. Rubino, Self-organizing Maps and Scaleinvariant Maps in Echo State Networks, 11th International Conference on Intelligent Systems Design and Applications, pp.94-99, 2011.

. Levenberg-marquardt, Training Algorithms for Random Neural Networks, Computer Journal, vol.54, issue.1, pp.125-135, 2011.

S. Basterrech and G. Rubino, Echo State Queueing Network: A new reservoir computing learning tool, 2013 IEEE 10th Consumer Communications and Networking Conference (CCNC), pp.118-123, 2013.
DOI : 10.1109/CCNC.2013.6488435

URL : https://hal.archives-ouvertes.fr/hal-00771627

O. Brun, L. Wang, and E. Gelenbe, Big Data for Autonomic Intercontinental Overlays, IEEE Journal on Selected Areas in Communications, vol.34, issue.3, pp.575-583, 2016.
DOI : 10.1109/JSAC.2016.2525518

URL : https://hal.archives-ouvertes.fr/hal-01461990

J. B. Butcher, D. Verstraeten, B. Schrauwen, C. R. Day, and P. W. Haycock, Reservoir computing and extreme learning machines for non-linear time-series data analysis, Neural Networks, vol.38, pp.76-89, 2013.
DOI : 10.1016/j.neunet.2012.11.011

K. Doya, Bifurcations in the learning of recurrent neural networks, [Proceedings] 1992 IEEE International Symposium on Circuits and Systems, pp.2777-2780, 1992.
DOI : 10.1109/ISCAS.1992.230622

J. L. Elman, Finding Structure in Time, Cognitive Science, vol.49, issue.2, pp.179-211, 1990.
DOI : 10.1007/BF00308682

A. A. Ferreira, T. B. Ludermir, R. R. De, and . Aquino, An approach to reservoir computing design and training, Expert Systems with Applications, vol.40, issue.10, pp.4172-4182, 2013.
DOI : 10.1016/j.eswa.2013.01.029

E. Gelenbe, Random Neural Networks with Negative and Positive Signals and Product Form Solution, Neural Computation, vol.1, issue.4, pp.502-510, 1989.
DOI : 10.1162/neco.1989.1.4.502

E. Gelenbe, Product-form queueing networks with negative and positive customers, Journal of Applied Probability, vol.28, issue.03, pp.656-663, 1991.
DOI : 10.1145/321879.321887

E. Gelenbe, Learning in the Recurrent Random Neural Network, Neural Computation, vol.85, issue.1, pp.154-511, 1993.
DOI : 10.1162/neco.1989.1.2.161

E. Gelenbe, The spiked random neural network: nonlinearity, learning and approximation, 1998 Fifth IEEE International Workshop on Cellular Neural Networks and their Applications. Proceedings (Cat. No.98TH8359), pp.14-19, 1998.
DOI : 10.1109/CNNA.1998.685674

E. Gelenbe, Z. Mao, and Y. Li, Function Approximation by Random Neural Networks with a Bounded Number of Layers, Journal of Differential Equations and Dynamical Systems, vol.12, pp.143-170, 2004.
DOI : 10.1142/9781860948923_0005

E. Gelenbe and M. Schassberger, Stability of Product Form G-Networks, Probability in the Engineering and Informational Sciences, pp.271-276, 1992.
DOI : 10.2307/3214756

E. Gelenbe, G-Networks with Signals and Batch Removal, Probability in the Engineering and Informational Sciences, pp.335-342, 1993.
DOI : 10.2307/3214499

E. Gelenbe, Learning in the Recurrent Random Neural Network, Neural Computation, vol.85, issue.1, pp.154-164, 1993.
DOI : 10.1162/neco.1989.1.2.161

E. Gelenbe, Steps toward self-aware networks, Communications of the ACM, vol.52, issue.7, pp.66-75, 2009.
DOI : 10.1145/1538788.1538809

URL : http://www.inf.brad.ac.uk/~iawan/aina/Gelenbe-CACM.pdf

E. Gelenbe and J. Fourneau, G-networks with resets. Performance Evaluation, pp.179-191, 2002.
DOI : 10.1016/s0166-5316(02)00127-x

E. Gelenbe, Y. Zhi-hong-mao, and L. Da, Function Approximation by Random Neural Networks with a Bounded Number of Layers, Journal of Differential Equations and Dynamical Systems, vol.12, issue.12, pp.143-170, 2004.
DOI : 10.1142/9781860948923_0005

E. Gelenbe and K. Hussain, Learning in the multiple class random neural network, IEEE Transactions on Neural Networks, vol.13, issue.6, pp.1257-1267, 2002.
DOI : 10.1109/TNN.2002.804228

E. Gelenbe, Z. Mao, and Y. Li, Function approximation with spiked random networks, IEEE Transactions on Neural Networks, vol.10, issue.1, pp.3-9, 1999.
DOI : 10.1109/72.737488

URL : http://www.cs.ucf.edu/csdept/faculty/papers/IEEETNN1.ps

E. Gelenbe and S. Timotheou, Random Neural Networks with Synchronized Interactions, Neural Computation, vol.70, issue.4, pp.2308-2324, 2008.
DOI : 10.1103/PhysRevE.64.011920

URL : http://sa.ee.ic.ac.uk/publications/RNN-SI.pdf

E. Gelenbe and Y. Yin, Deep learning with random neural networks, SAI Intelligent Systems Conference 2016. IEEEXpress, 2016.
DOI : 10.1109/ijcnn.2016.7727393

S. Hochreiter and J. Schmidhuber, Long Short-Term Memory, Neural Computation, vol.4, issue.8, pp.1735-1780, 1997.
DOI : 10.1016/0893-6080(88)90007-X

H. Jaeger, The " echo state " approach to analysing and training recurrent neural networks, 2001.

M. Jordan, Serial Order: A Parallel Distributed Processing Approach, 1986.
DOI : 10.1016/S0166-4115(97)80111-2

A. Likas and A. Stafylopatis, Training the random neural network using quasi-Newton methods, European Journal of Operational Research, vol.126, issue.2, pp.331-339, 2000.
DOI : 10.1016/S0377-2217(99)00482-8

M. Luko?-sevi?-cius, On self-organizing reservoirs and their hierarchies, 2010.

M. Luko?-sevi?-cius and H. Jaeger, Reservoir computing approaches to recurrent neural network training, Computer Science Review, vol.3, issue.3, pp.127-149, 2009.
DOI : 10.1016/j.cosrev.2009.03.005

W. Maass, Noisy spiking neurons with temporal coding have more computational power than sigmoidal neurons, 1999.

W. Maass, T. Natschläger, and H. Markram, Real-Time Computing Without Stable States: A New Framework for Neural Computation Based on Perturbations, Neural Computation, vol.7, issue.11, pp.2531-2560, 2002.
DOI : 10.1038/35009102

W. Maass, T. Natschläger, and H. Markram, Computational Models for Generic Cortical Microcircuits, Neuroscience Databases. A Practical Guide, pp.121-136, 2003.
DOI : 10.1201/9780203494462.ch18

S. Mohamed and G. Rubino, A study of real-time packet video quality using random neural networks, IEEE Transactions on Circuits and Systems for Video Technology, vol.12, issue.12, pp.1071-1083, 2002.
DOI : 10.1109/TCSVT.2002.806808

URL : https://hal.archives-ouvertes.fr/inria-00072063

R. Pascanu, T. Mikolov, and Y. Bengio, On the difficulty of training recurrent neural networks, Proceedings of the 30th International Conference on Machine Learning, pp.37-48, 2013.

H. Paugam-moisy and S. M. Bohte, Handbook of Natural Computing, chapter Computing with Spiking Neuron Networks, 2009.

W. Press, S. Teukolsky, W. Vetterling, and B. Flannery, Numerical Recipes in C, 1992.

N. Robinzonov, G. Tutz, and T. Hothorn, Boosting techniques for nonlinear time series models AStA Advances in Statistical Analysis, pp.99-122, 2012.

A. Rodan, P. Ti, and ?. No, Minimum Complexity Echo State Network, IEEE Transactions on Neural Networks, vol.22, issue.1, pp.131-144, 2011.
DOI : 10.1109/TNN.2010.2089641

G. Rubino and M. Varela, A new approach for the prediction of end-to-end performance of multimedia streams, First International Conference on the Quantitative Evaluation of Systems, 2004. QEST 2004. Proceedings., pp.110-119, 2004.
DOI : 10.1109/QEST.2004.1348025

Â. Gerardo and R. , Quantifying the quality of audio and video transmissions over the internet: The psqa approach Design and Operations of Communication Networks: A Review of Wired and Wireless Modelling and Management Challenges, 2005.

D. E. Rumelhart, G. E. Hinton, and J. L. Mcclelland, A general framework for parallel distributed processing of Computational Models of Cognition and Perception, Parallel Distributed Processing: Explorations in the Microstructure of Cognition, pp.45-76, 1986.

B. Schrauwen, M. Wardermann, D. Verstraeten, J. J. Steil, and D. Stroobandt, Improving reservoirs using intrinsic plasticity, Neurocomputing, vol.71, issue.7-9, pp.1159-1171, 2007.
DOI : 10.1016/j.neucom.2007.12.020

URL : http://escher.elis.ugent.be/publ/Edocs/doc.php?file=P108_028.pdf

C. Sheng, J. Zhao, W. Wang, and H. Leung, Prediction Intervals for a Noisy Nonlinear Time Series Based on a Bootstrapping Reservoir Computing Network Ensemble, IEEE Transactions on Neural Networks and Learning Systems, vol.24, issue.7, pp.1036-1048, 2013.
DOI : 10.1109/TNNLS.2013.2250299

T. Hava, E. D. Siegelmann, and . Sontag, Turing computability with neural nets, Applied Mathematics Letters, vol.4, issue.6, pp.77-80, 1991.

J. Jochen and . Steil, Backpropagation-Decorrelation: online recurrent learning with O(N) complexity, Proceedings of IJCNN'04, 2004.

S. Timotheou, The Random Neural Network: A Survey, The Computer Journal, vol.53, issue.3, pp.251-267, 2010.
DOI : 10.1093/comjnl/bxp032

D. Verstraeten, B. Schrauwen, M. D. Haene, and D. Stroobandt, An experimental unification of reservoir computing methods, Neural Networks, vol.20, issue.3, pp.287-289, 2007.
DOI : 10.1016/j.neunet.2007.04.003

G. Wainrib, N. Mathieu, and . Galtier, A local Echo State Property through the largest Lyapunov exponent, Neural Networks, vol.76, pp.39-45, 2016.
DOI : 10.1016/j.neunet.2015.12.013

B. Zhang, D. J. Miller, and Y. Wang, Nonlinear System Modeling With Random Matrices: Echo State Networks Revisited, IEEE Transactions on Neural Networks and Learning Systems, vol.23, issue.1, pp.39-45, 2016.
DOI : 10.1109/TNNLS.2011.2178562