Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 May 23:16:845955.
doi: 10.3389/fnbot.2022.845955. eCollection 2022.

Continual Sequence Modeling With Predictive Coding

Affiliations

Continual Sequence Modeling With Predictive Coding

Louis Annabi et al. Front Neurorobot. .

Abstract

Recurrent neural networks (RNNs) have been proved very successful at modeling sequential data such as language or motions. However, these successes rely on the use of the backpropagation through time (BPTT) algorithm, batch training, and the hypothesis that all the training data are available at the same time. In contrast, the field of developmental robotics aims at uncovering lifelong learning mechanisms that could allow embodied machines to learn and stabilize knowledge in continuously evolving environments. In this article, we investigate different RNN designs and learning methods, that we evaluate in a continual learning setting. The generative modeling task consists in learning to generate 20 continuous trajectories that are presented sequentially to the learning algorithms. Each method is evaluated according to the average prediction error over the 20 trajectories obtained after complete training. This study focuses on learning algorithms with low memory requirements, that do not need to store past information to update their parameters. Our experiments identify two approaches especially fit for this task: conceptors and predictive coding. We suggest combining these two mechanisms into a new proposed model that we label PC-Conceptors that outperforms the other methods presented in this study.

Keywords: Reservoir Computing (RC); conceptors; continual learning; predictive coding; recurrent neural networks (RNN).

PubMed Disclaimer

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Figures

Figure 1
Figure 1
Three body configurations taken from a trajectory capturing a jump motion.
Figure 2
Figure 2
Simple RNN model.
Figure 3
Figure 3
PC-RNN-V model.
Figure 4
Figure 4
PC-RNN-HC model.
Figure 5
Figure 5
Score estimation of the hyperparameter optimizer with regard to the learning rate λ and the coefficient β, for the EWC model.
Figure 6
Figure 6
Continual learning results with the ESN model (left) and the Conceptors model (right). We represent the average prediction error over 10 seeds, for the continual learning of 20 sequential patterns, obtained on the first test set. The colored lines correspond to the prediction error on each individual task, and the black line corresponds to the prediction error averaged on all tasks. The 20 tasks are delimited by the dashed gray lines.
Figure 7
Figure 7
Comparison between the four learning methods for the output weights on the first test set. The 20 tasks are delimited by the dashed gray lines.
Figure 8
Figure 8
Comparison between the three learning methods for the input weights. The PC-RNN-V model, where no learning is performed on the input weights, is also displayed as a baseline. The 20 tasks are delimited by the dashed gray lines.
Figure 9
Figure 9
PC-Conceptors model.
Figure 10
Figure 10
Continual learning results using the PC-Conceptors. We represent the average prediction error over 10 seeds, for the continual learning of 20 sequential patterns, using the PC-RNN-HC-A model with Conceptors. The colored lines correspond to the prediction error on each individual task, and the black line corresponds to the prediction error averaged on all tasks. The 20 tasks are delimited by the dashed gray lines.
Figure 11
Figure 11
Comparison between the ESN, the Conceptors model, the PC-RNN-HC-A model, and the PC-Conceptors model on the first test set. The 20 tasks are delimited by the dashed gray lines.

References

    1. Annabi L., Pitti A., Quoy M. (2021a). Bidirectional interaction between visual and motor generative models using predictive coding and active inference. Neural Netw. 143, 638–656. 10.1016/j.neunet.2021.07.016 - DOI - PubMed
    1. Annabi L., Pitti A., Quoy M. (2021b). “A predictive coding account for chaotic itinerancy,” in Artificial Neural Networks and Machine Learning-ICANN 2021, eds I. Farkaš, P. Masulli, S. Otte, and S. Wermter (Cham: Springer International Publishing: ), 581–592.
    1. Clark A.. (2013). Whatever next? predictive brains, situated agents, and the future of cognitive science. Behav. Brain Sci. 36, 181–204. 10.1017/S0140525X12000477 - DOI - PubMed
    1. Collins J., Sohl-Dickstein J., Sussillo D. (2016). Capacity and trainability in recurrent neural networks. stat 1050:29. - PubMed
    1. Cossu A., Bacciu D., Carta A., Gallicchio C., Lomonaco V. (2021a). Continual learning with echo state networks. arXiv preprint arXiv:2105.07674. 10.14428/esann/2021.ES2021-80 - DOI