Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Journal articles
  4. Learning in Volatile Environments With the Bayes Factor Surprise
 
research article

Learning in Volatile Environments With the Bayes Factor Surprise

Liakoni, Vasiliki  
•
Modirshanechi, Alireza  
•
Gerstner, Wulfram  
Show more
February 1, 2021
Neural Computation

Surprise-based learning allows agents to rapidly adapt to nonstationary stochastic environments characterized by sudden changes. We show that exact Bayesian inference in a hierarchical model gives rise to a surprise-modulated trade-off between forgetting old observations and integrating them with the new ones. The modulation depends on a probability ratio, which we call the Bayes Factor Surprise, that tests the prior belief against the current belief. We demonstrate that in several existing approximate algorithms, the Bayes Factor Surprise modulates the rate of adaptation to new observations. We derive three novel surprise-based algorithms, one in the family of particle filters, one in the family of variational learning, and one in the family of message passing, that have constant scaling in observation sequence length and particularly simple update dynamics for any distribution in the exponential family. Empirical results show that these surprise-based algorithms estimate parameters better than alternative approximate approaches and reach levels of performance comparable to computationally more expensive algorithms. The Bayes Factor Surprise is related to but different from the Shannon Surprise. In two hypothetical experiments, we make testable predictions for physiological indicators that dissociate the Bayes Factor Surprise from the Shannon Surprise. The theoretical insight of casting various approaches as surprise-based learning, as well as the proposed online algorithms, may be applied to the analysis of animal and human behavior and to reinforcement learning in nonstationary environments.

  • Details
  • Metrics
Type
research article
DOI
10.1162/neco_a_01352
Web of Science ID

WOS:000613887400001

Author(s)
Liakoni, Vasiliki  
Modirshanechi, Alireza  
Gerstner, Wulfram  
Brea, Johanni  
Date Issued

2021-02-01

Publisher

MIT PRESS

Published in
Neural Computation
Volume

33

Issue

2

Start page

269

End page

340

Subjects

Computer Science, Artificial Intelligence

•

Neurosciences

•

Computer Science

•

Neurosciences & Neurology

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
LCN  
Available on Infoscience
April 10, 2021
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/177076
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés