Skip to Main content Skip to Navigation
Poster communications

Recurrent Neural Networks Models for Developmental Language Acquisition: Reservoirs Outperform LSTMs

Xavier Hinaut 1 Alexandre Variengien 2, 1
1 Mnemosyne - Mnemonic Synergy
LaBRI - Laboratoire Bordelais de Recherche en Informatique, Inria Bordeaux - Sud-Ouest, IMN - Institut des Maladies Neurodégénératives [Bordeaux]
Abstract : We previously developed cortico-striatal models for sentence comprehension (Hinaut & Dominey 2013) and sentence production (Hinaut et al. 2015). The sentence comprehension model is based on the reservoir computing principle: a random recurrent neural network (a reservoir) provides a rich recombination of sequential word inputs (e.g. a piece of prefrontal cortex), and an output layer (e.g. striatum) learns to "reads- out" the roles of words in the sentence from the internal recurrent dynamics. The model has several interesting properties, like the ability to predict the semantic roles of words during online processing. Additionally, we demonstrated its robustness to various corpus complexities, in different languages, and even its ability to work with bilingual inputs. In this study, we propose to (1) use the model in a new task related to a developmental language acquisition (i.e. Cross-Situational Learning), (2) provide a quantitative comparison with one of the best performing neural networks for sequential tasks (a LSTM), and (3) provide a qualitative analysis on the way reservoirs and LSTMs solve the task. This new Cross-Situational Task is as follows: for a given sentence, the target output provided often contains more detailed features than what is available in the sentence. Thus, the models have not only to learn how to parse sentences to extract useful information, but also to statistically infer which word is associated with which feature. While reservoir units are modelled as leaky average firing rate neurons, LSTM units are engineered to gate information using a costly and biologically implausible learning algorithm (Back-Propagation Through Time). We found that both models were able to successfully learn the task: the LSTM reached slightly better performance for the basic corpus, but the reservoir was able to significantly outperform LSTMs on more challenging corpora with increasing vocabulary sizes (for a given set of hyperparameters). We analyzed the hidden activations of internal units of both models. Despite the deep differences between both models (trained or fixed internal weights), we were able to uncover similar inner dynamics: the most useful units (with strongest weights to the output layer) seemed tuned to keep track of several specific words in the sentence. Because of its learning algorithm, it is predictable to see such behavior in a LTSM but not in a reservoir; in fact, the LSTM contained more tuned-like units than the reservoir. These differences between LSTMs and reservoirs highlights differences between classical Deep Learning approaches (based on back-propagation algorithm) and more plausible brain learning mechanisms. First, the reservoir is more efficient in terms of training time and cost (the LSTM needs several passes on the training data, while the reservoir uses it only one). Secondly, only the reservoir model seems to scale to larger corpora without the need to adapt specifically the hyperparameters of the model. Finally, the presence of more tuned units in the LSTM compared to the reservoir might be an explanation of why the LSTM seems to overfit too much to training data and have limited generalization capabilities when the learning data available becomes limited.
Complete list of metadata

https://hal.inria.fr/hal-03146558
Contributor : Xavier Hinaut Connect in order to contact the contributor
Submitted on : Friday, February 19, 2021 - 5:03:55 PM
Last modification on : Friday, January 21, 2022 - 3:10:39 AM
Long-term archiving on: : Thursday, May 20, 2021 - 6:28:19 PM

File

HinautVariengien_SNL2020_poste...
Files produced by the author(s)

Identifiers

  • HAL Id : hal-03146558, version 1

Collections

Citation

Xavier Hinaut, Alexandre Variengien. Recurrent Neural Networks Models for Developmental Language Acquisition: Reservoirs Outperform LSTMs. SNL 2020 - 12th Annual Meeting of the Society for the Neurobiology of Language, Oct 2020, Virtual Edition, Canada. ⟨hal-03146558⟩

Share

Metrics

Les métriques sont temporairement indisponibles