4.5 Article

Learning in Volatile Environments With the Bayes Factor Surprise

Journal

NEURAL COMPUTATION
Volume 33, Issue 2, Pages 269-340

Publisher

MIT PRESS
DOI: 10.1162/neco_a_01352

Keywords

-

Funding

  1. Swiss National Science Foundation [200020 184615]
  2. European Union [785907]

Ask authors/readers for more resources

Surprise-based learning allows agents to adapt rapidly in nonstationary environments, with the Bayes Factor Surprise modulating the balance between forgetting old observations and integrating new ones. Novel surprise-based algorithms have constant scaling in observation sequence length and simple update dynamics for distributions in the exponential family, showing better parameter estimation compared to alternative approaches. Theoretical insight and proposed online algorithms in the study may be useful for analyzing behavior in animals and humans, as well as reinforcement learning in changing environments.
Surprise-based learning allows agents to rapidly adapt to nonstationary stochastic environments characterized by sudden changes. We show that exact Bayesian inference in a hierarchical model gives rise to a surprise-modulated trade-off between forgetting old observations and integrating them with the new ones. The modulation depends on a probability ratio, which we call the Bayes Factor Surprise, that tests the prior belief against the current belief. We demonstrate that in several existing approximate algorithms, the Bayes Factor Surprise modulates the rate of adaptation to new observations. We derive three novel surprise-based algorithms, one in the family of particle filters, one in the family of variational learning, and one in the family of message passing, that have constant scaling in observation sequence length and particularly simple update dynamics for any distribution in the exponential family. Empirical results show that these surprise-based algorithms estimate parameters better than alternative approximate approaches and reach levels of performance comparable to computationally more expensive algorithms. The Bayes Factor Surprise is related to but different from the Shannon Surprise. In two hypothetical experiments, we make testable predictions for physiological indicators that dissociate the Bayes Factor Surprise from the Shannon Surprise. The theoretical insight of casting various approaches as surprise-based learning, as well as the proposed online algorithms, may be applied to the analysis of animal and human behavior and to reinforcement learning in nonstationary environments.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

Article Computer Science, Artificial Intelligence

Exponentially Long Orbits in Hopfield Neural Networks

Samuel P. Muscinelli, Wulfram Gerstner, Johanni Brea

NEURAL COMPUTATION (2017)

Review Neurosciences

Eligibility Traces and Plasticity on Behavioral Time Scales: Experimental Support of NeoHebbian Three-Factor Learning Rules

Wulfram Gerstner, Marco Lehmann, Vasiliki Liakoni, Dane Corneil, Johanni Brea

FRONTIERS IN NEURAL CIRCUITS (2018)

Article Neurosciences

Trial-by-trial surprise-decoding model for visual and auditory binary oddball tasks

Alireza Modirshanechi, Mohammad Mahdi Kiani, Hamid Aghajan

NEUROIMAGE (2019)

Article Computer Science, Artificial Intelligence

Biologically plausible deep learning - But how far can we go with shallow networks?

Bernd Illing, Wulfram Gerstner, Johanni Brea

NEURAL NETWORKS (2019)

Article Multidisciplinary Sciences

Testing two competing hypotheses for Eurasian jays' caching for the future

Piero Amodio, Johanni Brea, Benjamin G. Farrar, Ljerka Ostojic, Nicola S. Clayton

Summary: The study explored prospective caching behaviors in corvids and found that neither the Compensatory Caching Hypothesis nor the Future Planning Hypothesis were supported by the experimental results. Instead, the best explanation was a uniform distribution of food caches across different caching locations.

SCIENTIFIC REPORTS (2021)

Article Neurosciences

Rapid suppression and sustained activation of distinct cortical regions for a delayed sensory-triggered motor response

Vahid Esmaeili, Keita Tamura, Samuel P. Muscinelli, Alireza Modirshanechi, Marta Boscaglia, Ashley B. Lee, Anastasiia Oryshchuk, Georgios Foustoukos, Yanqi Liu, Sylvain Crochet, Wulfram Gerstner, Carl C. H. Petersen

Summary: This study used a combination of calcium imaging, electrophysiology, and optogenetics to track the cortical activity sequence in mice from whisker stimulation to delayed licking. It found that enhanced activity in the secondary whisker motor cortex and transient reduction in orofacial sensorimotor cortex played important roles in completing the task, while sustained activity in the frontal cortex was crucial for licking during the response period.

NEURON (2021)

Article Biochemical Research Methods

Novelty is not surprise: Human exploratory and adaptive behavior in sequential decision-making

He A. Xu, Alireza Modirshanechi, Marco P. Lehmann, Wulfram Gerstner, Michael H. Herzog

Summary: This research demonstrates the importance of incorporating surprise and novelty into reinforcement learning theories to explain human behavior. It is found that human decisions are mainly influenced by model-free action choices, with the world-model playing a role in detecting surprising events.

PLOS COMPUTATIONAL BIOLOGY (2021)

Article Neurosciences

Brain signals of a Surprise-Actor-Critic model: Evidence for multiple learning modules in human decision making

Vasiliki Liakoni, Marco P. Lehmann, Alireza Modirshanechi, Johanni Brea, Antoine Lutti, Wulfram Gerstner, Kerstin Preuschoff

Summary: Humans have a remarkable capability of learning how to reach a reward over long series of actions, potentially guided by multiple parallel learning modules. This study introduces a complex decision-making task and tests various learning algorithms to identify model-free and surprise-based learning signals in the brain.

NEUROIMAGE (2022)

Article Mathematics, Interdisciplinary Applications

A taxonomy of surprise definitions

Alireza Modirshanechi, Johanni Brea, Wulfram Gerstner

Summary: Surprising events have a measurable impact on brain activity and human behavior, affecting learning, memory, and decision-making. The definition of surprise lacks consensus. In this study, the authors identify and classify 18 mathematical definitions of surprise, proposing a taxonomy based on the measured quantity. This research lays the foundation for studying the functional roles and physiological signatures of surprise in the brain.

JOURNAL OF MATHEMATICAL PSYCHOLOGY (2022)

Article Multidisciplinary Sciences

Computational models of episodic-like memory in food-caching birds

Johanni Brea, Nicola S. Clayton, Wulfram Gerstner

Summary: In this study, the authors present computational models of neural networks to explain how the 'what', 'where', and 'when' of past experiences are stored and retrieved in episodic memories for suitable decisions, using food-caching birds as examples. They propose a model and neural implementation of food-caching behavior, suggesting that associative reinforcement learning without mental time-travel is sufficient to explain the results of behavioral experiments with food-caching birds.

NATURE COMMUNICATIONS (2023)

Article Neurosciences

Surprise and novelty in the brain

Alireza Modirshanechi, Sophia Becker, Johanni Brea, Wulfram Gerstner

Summary: This article systematically investigates the relationship between different aspects of surprise and novelty with brain functions and physiological signals, addressing concerns about the existence of same functionalities and mechanisms in the brain. It emphasizes the significance of computational modeling and quantifiable definitions in providing novel interpretations of previous findings.

CURRENT OPINION IN NEUROBIOLOGY (2023)

Article Computer Science, Interdisciplinary Applications

GaussianProcesses.jl: A Nonparametric Bayes Package for the Julia Language

Jamie Fairbrother, Christopher Nemeth, Maxime Rischard, Johanni Brea, Thomas Pinder

Summary: This paper introduces the GaussianProcesses.jl package developed for the Julia programming language, utilizing the computational benefits of Julia for fast, flexible and user-friendly Gaussian process modeling. The package provides various mean and kernel functions, inference tools, alternative likelihood functions for non-Gaussian data, and sparse approximations.

JOURNAL OF STATISTICAL SOFTWARE (2022)

No Data Available