Resource type
Thesis type
(Thesis) M.Sc.
Date created
2021-09-01
Authors/Contributors
Author: Salari, Mohammad Hadi
Abstract
Neural networks have been successful on many individual tasks. However, they work poorly on a sequential stream of tasks when they do not have access to the previous data. This problem is called Catastrophic Forgetting and recently has been studied in the field of Continual Learning. In this thesis, we propose four Continual Learning methods which approximate the loss function on previous tasks by a second-order Taylor approximation and use them as regularizers to maintain the performance on previous tasks without any access to previous data. To do that, we use a Kronecker-factored Hessian approximation to make the training process memory and computationally efficient. We evaluate our methods on two Domain Incremental datasets called Permuted MNIST and Rotated MNIST and investigate the performance and efficiency trade-off on them.
Document
Extent
28 pages.
Identifier
etd21634
Copyright statement
Copyright is held by the author(s).
Supervisor or Senior Supervisor
Thesis advisor: Mori, Greg
Language
English
Member of collection
Download file | Size |
---|---|
etd21634.pdf | 2.68 MB |