Tracking emotions: intrinsic motivation grounded on multi-level prediction error dynamics

Happy that my paper on “Tracking emotions: intrinsic motivation grounded on multi-level prediction error dynamics”, co-authored with Alejandra Ciria (UNAM, MX) and Bruno Lara (UAEM, MX), has been accepted for presentation at IEEE ICDL-Epirob 2020!

In this work, we propose a learning architecture that generates exploratory behaviours towards self-generated goals in a simulated robot, and that regulates goal selection and the balance between exploitation and exploration through a multi-level monitoring of prediction error dynamics.

The system is made of: 1) a convolutional autoencoder for unsupervised learning of low-dimensional features from visual inputs; 2) a self-organising map for online learning of visual goals; two deep neural networks, trained in an online fashion, encoding controller and predictor of the system. Memory replay is employed to face catastrophic forgetting issues.

A multi-level monitoring mechanism keeps track of two errors: (1) a high-level, general error of the system, i.e. MSE of the forward model calculated on a test dataset; (2) low-level goal errors, i.e. the prediction errors estimated when trying to reach each specific goal.

The system maintains a buffer of high-level MSE observed during a specific time window. After every update of the MSE buffer, a linear regression is calculated on the stored values over time, whose slope indicates the trend of the general error of the system.

This trend modulates computational resources (size of goal error buffers) and exploration noise: when overall performances improve, the necessity of tracking the goal error dynamics is reduced. On the contrary, the system widens the time window on which goal errors are monitored.

We discuss the tight relationship that PE dynamics may have with the emotional valence of action. PE dynamics may be fundamental cause of emotional valence of action: positive valence linked to an active reduction of PE and a negative valence to a continuous increase of PE.

Read the full paper here: https://arxiv.org/abs/2007.14632 (pre-print)!

Prediction-error driven memory consolidation for continual learning and adaptive greenhouse models

Check my AI Transfer work submitted to Springer KI (German Journal on Artificial Intelligence, special issue on Developmental Robotics) on “Prediction error-driven memory consolidation for continual learning”, applied on data from innovative greenhouses: https://arxiv.org/abs/2006.12616.

Episodic memory replay and prediction-error driven consolidation are used to tackle online learning in deep recurrent neural networks. Inspired by evidences from cognitive sciences and neuroscience, memories are retained depending on their congruency with prior knowledge.
This congruency is estimated in terms of prediction errors resulting from a generative model. In particular, our framework chooses which samples to maintain in the episodic memory based on their expected contribution to the learning progress.
Different retention strategies are compared. We analyse their impact on the variance of the samples stored in the memory and on the stability/plasticity of the model.

Co-authored with Luis Miranda and Uwe Schmidt, Humboldt-Universität zu Berlin.

SAGE Adaptive Behavior

Our article on “Intrinsic Motivation and Episodic Memories for Robot Exploration of High-Dimensional Sensory Spaces” is out in SAGE Adaptive Behaviour! Pre-print available here: https://arxiv.org/abs/2001.01982

Co-authored with Antonio Pico (HU-Berlin), Verena Hafner (HU-Berlin), Peter Hanappe (Sony CSL), David Colliaux (Sony CSL) and Timothee Wintz (Sony CSL)