M. D. Crawford, G. J. Brown, M. P. Cooke, and P. D. Green, Design, collection and analysis of a multi-simultaneous-speaker corpus, Proceedings of the IOA, 1994.

K. Tanaka, S. Hayamizu, Y. Yamashita, K. Shikano, S. Itahashi et al., Design and data collection for a spoken dialog database in the Real World Computing (RWC) program, The Journal of the Acoustical Society of America, vol.100, issue.4, 1996.
DOI : 10.1121/1.416338

H. Hirsch and D. Pearce, The Aurora experimental framework for the performance evaluation of speech recognition systems under noisy conditions, Proc. ASR2000, 2000.

T. H. Crystal, A. Schmidt-nielsen, and E. Marsh, Speech in noisy environments (spine) adds new dimension to speech recognition R&D, Proceedings of the second international conference on Human Language Technology Research -, 2002.
DOI : 10.3115/1289189.1289199

K. Tanaka, K. Itou, M. Ihara, and R. Oka, Constructing a meeting speech corpus, IPSJ Tech. Rep, 2001.

S. Nakamura, K. Hiyane, F. Asano, T. Nishiura, and T. Yamada, Acoustical sound database in real environments for sound scene understanding and hands-free speech recognition, Proc. LREC, 2000.

A. Moreno, B. Lindberg, C. Draxler, G. Richard, and K. Choukri, SPEECHDAT-CAR. a large speech database for automotive environments, Proc. LREC, 2000.

N. Parihar and J. Picone, DSR front-end large vocabulary continuous speech recognition evaluation, 2002.

L. Lamel, F. Schiel, A. Fourcin, J. Mariani, and H. Tillman, The translingual English database (TED), Proc. ICSLP, 1994.

E. K. Patterson, S. Gurbuz, Z. Tufekci, and J. N. Gowdy, Moving-Talker, Speaker-Independent Feature Study, and Baseline Results Using the CUAVE Multimodal Speech Corpus, EURASIP Journal on Advances in Signal Processing, vol.2002, issue.11, 2002.
DOI : 10.1155/S1110865702206101

J. H. Hansen, P. Angkititrakul, J. Plucienkowski, S. Gallant, and U. , CU-Move " : Analysis & corpus development for interactive in-vehicle speech systems, Proc. Eurospeech, 2001.

S. Nakamura, K. Takeda, K. Yamamoto, T. Yamada, and S. Kuroiwa, AURORA-2J: An Evaluation Framework for Japanese Noisy Speech Recognition, IEICE Transactions on Information and Systems, vol.88, issue.3, 2005.
DOI : 10.1093/ietisy/e88-d.3.535

B. Lee, M. Hasegawa-johnson, C. Goudeseune, S. Kamdar, and S. Borys, AVICAR: audio-visual speech corpus in a car environment, Proc. Interspeech, 2004.

G. Lathoud, J. Odobez, and D. Gatica-perez, AV16.3: An Audio-Visual Corpus for Speaker Localization and Tracking, Proc. MLMI, 2004.
DOI : 10.1007/978-3-540-30568-2_16

A. Janin, D. Baron, J. Edwards, D. Ellis, and D. Gelbart, The ICSI Meeting Corpus, 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP '03)., 2003.
DOI : 10.1109/ICASSP.2003.1198793

J. S. Garofolo, C. D. Laprun, M. Michel, V. M. Stanford, and E. Tabassi, The NIST meeting room pilot corpus, Proc. LREC, 2004.

D. Mostefa, N. Moreau, K. Choukri, G. Potamianos, and S. Chu, The CHIL audiovisual corpus for lecture and meeting analysis inside smart rooms, Language Resources and Evaluation, vol.41, issue.3-4, pp.3-4, 2007.
DOI : 10.1007/s10579-007-9054-4

D. J. Iskra, B. Grosskopf, K. Marasek, H. Van-den-heuvel, F. Diehl et al., SPEECON speech databases for consumer devices: Database specification and validation, Proc. LREC, 2002.

S. Nakamura, M. Fujimoto, and K. Takeda, CENSREC2: Corpus and evaluation environments for in car continuous digit speech recognition, Proc. Interspeech, 2006.

M. Fujimoto, K. Takeda, and S. Nakamura, CENSREC-3: An Evaluation Framework for Japanese Speech Recognition in Real Car-Driving Environments, IEICE Transactions on Information and Systems, vol.89, issue.11, 2006.
DOI : 10.1093/ietisy/e89-d.11.2783

H. G. Hirsch, Aurora-5 experimental framework for the performance evaluation of speech recognition in case of a hands-free speech input in noisy environments, Niederrhein University of Applied Sciences , Tech. Rep, issue.21, 2007.

S. Renals, T. Hain, and H. Bourlard, Interpretation of Multiparty Meetings the AMI and Amida Projects, 2008 Hands-Free Speech Communication and Microphone Arrays, 2008.
DOI : 10.1109/HSCMA.2008.4538700

M. P. Cooke, J. R. Hershey, and S. J. Rennie, Monaural speech separation and recognition challenge, Computer Speech & Language, vol.24, issue.1, 2010.
DOI : 10.1016/j.csl.2009.02.006

URL : https://hal.archives-ouvertes.fr/hal-00598185

J. Segura, T. Ehrette, A. Potamianos, D. Fohr, I. Illina et al., The HIWIRE database, a noisy and non-native English speech corpus for cockpit communication, Tech. Rep, 2007.

Y. Hu and P. C. Loizou, Subjective comparison and evaluation of speech enhancement algorithms, Speech Communication, vol.49, issue.7-8, pp.7-8, 2007.
DOI : 10.1016/j.specom.2006.12.006

P. Angkititrakul, J. H. Hansen, S. Choi, T. Creek, and J. Hayes, UTDrive: The smart vehicle project, " in In-vehicle corpus and signal processing for driver behavior, 2009.

E. Vincent, S. Araki, F. J. Theis, G. Nolte, and P. Bofill, The signal separation evaluation campaign (2007???2010): Achievements and remaining challenges, Signal Processing, vol.92, issue.8, 2012.
DOI : 10.1016/j.sigpro.2011.10.007

URL : https://hal.archives-ouvertes.fr/inria-00579398

M. Lincoln, I. Mccowan, J. Vepa, and H. K. Maganti, The multi-channel Wall Street Journal audio visual corpus (MC-WSJ-AV): specification and initial experiments, IEEE Workshop on Automatic Speech Recognition and Understanding, 2005., 2005.
DOI : 10.1109/ASRU.2005.1566470

E. Zwyssig, F. Faubel, S. Renals, and M. Lincoln, Recognition of overlapping speech using digital MEMS microphone arrays, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, 2013.
DOI : 10.1109/ICASSP.2013.6639033

T. Nishiura, M. Nakayama, Y. Denda, N. Kitaoka, and K. Yamamoto, Evaluation framework for distant-talking speech recognition under reverberant environments ? newest part of the CENSREC series ?, Proc. LREC, 2008.

A. Brutti, L. Cristoforetti, W. Kellermann, L. Marquardt, and M. Omologo, WOZ acoustic data collection for interactive TV, Proc. LREC, 2008.
DOI : 10.1007/s10579-010-9116-x

URL : http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.381.9336

A. Stupakov, E. Hanusa, D. Vijaywargi, D. Fox, and J. Bilmes, The design and collection of COSINE, a multi-microphone in situ speech corpus recorded in noisy environments, Computer Speech & Language, vol.26, issue.1, 2011.
DOI : 10.1016/j.csl.2010.12.003

E. Vincent, J. Barker, S. Watanabe, J. Le-roux, F. Nesta et al., The second ‘chime’ speech separation and recognition challenge: Datasets, tasks and baselines, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, 2013.
DOI : 10.1109/ICASSP.2013.6637622

G. Gravier, G. Adda, N. Paulsson, M. Carré, A. Giraudel et al., The ETAPE corpus for the evaluation of speech-based TV content processing in the French language, Proc. LREC, 2012.
URL : https://hal.archives-ouvertes.fr/hal-00712591

K. Kinoshita, M. Delcroix, T. Yoshioka, T. Nakatani, and E. Habets, The reverb challenge: A common evaluation framework for dereverberation and recognition of reverberant speech, 2013 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, 2013.
DOI : 10.1109/WASPAA.2013.6701894

C. Fox, Y. Liu, E. Zwyssig, and T. Hain, The Sheffield wargames corpus, Proc. Interspeech, 2013.

L. Cristoforetti, M. Ravanelli, M. Omologo, A. Sosi, A. Abad et al., The DIRHA simulated corpus, Proc. LREC, 2014.

N. Kawaguchi, S. Matsubara, Y. Yamaguchi, K. Takeda, and F. Itakura, Ciair in-car speech database, Proc. Interspeech, 2004.

D. Ellis, H. Satoh, and Z. Chen, Detecting proximity from personal audio recordings, Proc. Interspeech, 2014.

E. Zwyssig, Speech processing using digital mems microphones, 2013.