Abstract
Continual learning aims to enable machine learning models to learn a general solution space for past and future tasks in a sequential manner. Conventional models tend to forget the knowledge of previous tasks while learning a new task, a phenomenon known as catastrophic forgetting. When using Bayesian models in continual learning, knowledge from previous tasks can be retained in two ways: 1). posterior distributions over the parameters, containing the knowledge gained from inference in previous tasks, which then serve as the priors for the following task; 2). coresets, containing knowledge of data distributions of previous tasks. Here, we show that Bayesian continual learning can be facilitated in terms of these two means through the use of natural gradients and Stein gradients respectively.
Original language | English |
---|---|
Number of pages | 8 |
Publication status | Published - 24 Apr 2019 |
Event | Continual Learning Workshop of 32nd Conference on Neural Information Processing Systems (NeurIPS 2018) - Montreal, Quebec, Canada Duration: 2 Dec 2018 → 8 Dec 2018 Conference number: 32 https://nips.cc/Conferences/2018 |
Workshop
Workshop | Continual Learning Workshop of 32nd Conference on Neural Information Processing Systems (NeurIPS 2018) |
---|---|
Abbreviated title | NeurIPS 2018 |
Country/Territory | Canada |
City | Montreal, Quebec |
Period | 2/12/18 → 8/12/18 |
Internet address |
Bibliographical note
Presented at the NeurIPS 2018 workshop on Continual Learning https://sites.google.com/view/continual2018/homeKeywords
- cs.LG
- cs.AI
- stat.ML
Fingerprint
Dive into the research topics of 'Facilitating Bayesian Continual Learning by Natural Gradients and Stein Gradients'. Together they form a unique fingerprint.Student theses
-
Efficient Continual Learning: Approaches and Measures
Author: Chen, Y., 28 Sep 2021Supervisor: Flach, P. (Supervisor) & Diethe, T. (Supervisor)
Student thesis: Doctoral Thesis › Doctor of Philosophy (PhD)
File