Editing a Special Issue on “Prediction and Perception in Humans and Robots” IEEE TCDS

A call for papers for a Special Issue on “Prediction and Perception in Humans and Robots” I am editing in IEEE Transactions on Cognitive and Developmental Systems is open.  Call available here: IEEE TCDS – special issue.

Experimental research, robotics implementations, and interdisciplinary works are particularly welcome, as well as theoretical contributions in the form of original research articles, reviews, and commentaries. This special issue invites researchers investigating topics related, but not limited, to:

  • Attentional and gating mechanisms for action and sensory
    information processing
  • Context- and task-dependent perceptual optimization
  • Sensory attenuation/cancellation and sensory enhancement/facilitation
  • Interplay of predictive and attentional mechanisms for prediction error minimization
  • Multimodal integration and/or cross-modal interactions
  • State representation learning
  • Constrained innate priors that drive learning
  • Prediction in language learning and comprehension
  • Prediction error dynamics monitoring
  • Dreaming, non-conscious perception, hallucination, altered perceptual phenomena

IMPORTANT DATES

  • Paper submission deadline – 15 July 2021
  • Notification to authors – 15 September 2021
  • Deadline revised papers submission – 15 November 2021
  • Final version – 15 December 2021

Tracking emotions: intrinsic motivation grounded on multi-level prediction error dynamics

Happy that my paper on “Tracking emotions: intrinsic motivation grounded on multi-level prediction error dynamics”, co-authored with Alejandra Ciria (UNAM, MX) and Bruno Lara (UAEM, MX), has been accepted for presentation at IEEE ICDL-Epirob 2020!

In this work, we propose a learning architecture that generates exploratory behaviours towards self-generated goals in a simulated robot, and that regulates goal selection and the balance between exploitation and exploration through a multi-level monitoring of prediction error dynamics.

The system is made of: 1) a convolutional autoencoder for unsupervised learning of low-dimensional features from visual inputs; 2) a self-organising map for online learning of visual goals; two deep neural networks, trained in an online fashion, encoding controller and predictor of the system. Memory replay is employed to face catastrophic forgetting issues.

A multi-level monitoring mechanism keeps track of two errors: (1) a high-level, general error of the system, i.e. MSE of the forward model calculated on a test dataset; (2) low-level goal errors, i.e. the prediction errors estimated when trying to reach each specific goal.

The system maintains a buffer of high-level MSE observed during a specific time window. After every update of the MSE buffer, a linear regression is calculated on the stored values over time, whose slope indicates the trend of the general error of the system.

This trend modulates computational resources (size of goal error buffers) and exploration noise: when overall performances improve, the necessity of tracking the goal error dynamics is reduced. On the contrary, the system widens the time window on which goal errors are monitored.

We discuss the tight relationship that PE dynamics may have with the emotional valence of action. PE dynamics may be fundamental cause of emotional valence of action: positive valence linked to an active reduction of PE and a negative valence to a continuous increase of PE.

Read the full paper here: https://arxiv.org/abs/2007.14632 (pre-print)!

Prediction-error driven memory consolidation for continual learning and adaptive greenhouse models

Check my AI Transfer work submitted to Springer KI (German Journal on Artificial Intelligence, special issue on Developmental Robotics) on “Prediction error-driven memory consolidation for continual learning”, applied on data from innovative greenhouses: https://arxiv.org/abs/2006.12616.

Episodic memory replay and prediction-error driven consolidation are used to tackle online learning in deep recurrent neural networks. Inspired by evidences from cognitive sciences and neuroscience, memories are retained depending on their congruency with prior knowledge.
This congruency is estimated in terms of prediction errors resulting from a generative model. In particular, our framework chooses which samples to maintain in the episodic memory based on their expected contribution to the learning progress.
Different retention strategies are compared. We analyse their impact on the variance of the samples stored in the memory and on the stability/plasticity of the model.

Co-authored with Luis Miranda and Uwe Schmidt, Humboldt-Universität zu Berlin.