Generalisation guarantees for continual learning with orthogonal gradient descent
In Continual Learning settings, deep neural networks are prone to Catastrophic Forgetting.
Orthogonal Gradient Descent was proposed to tackle the challenge. However, no theoretical
guarantees have been proven yet. We present a theoretical framework to study Continual
Learning algorithms in the Neural Tangent Kernel regime. This framework comprises closed
form expression of the model through tasks and proxies for Transfer Learning,
generalisation and tasks similarity. In this framework, we prove that OGD is robust to …
Orthogonal Gradient Descent was proposed to tackle the challenge. However, no theoretical
guarantees have been proven yet. We present a theoretical framework to study Continual
Learning algorithms in the Neural Tangent Kernel regime. This framework comprises closed
form expression of the model through tasks and proxies for Transfer Learning,
generalisation and tasks similarity. In this framework, we prove that OGD is robust to …
Generalisation guarantees for continual learning with orthogonal gradient descent
M Abbana Bennani, T Doan, M Sugiyama - arXiv e-prints, 2020 - ui.adsabs.harvard.edu
Abstract In Continual Learning settings, deep neural networks are prone to Catastrophic
Forgetting. Orthogonal Gradient Descent was proposed to tackle the challenge. However, no
theoretical guarantees have been proven yet. We present a theoretical framework to study
Continual Learning algorithms in the Neural Tangent Kernel regime. This framework
comprises closed form expression of the model through tasks and proxies for Transfer
Learning, generalisation and tasks similarity. In this framework, we prove that OGD is robust …
Forgetting. Orthogonal Gradient Descent was proposed to tackle the challenge. However, no
theoretical guarantees have been proven yet. We present a theoretical framework to study
Continual Learning algorithms in the Neural Tangent Kernel regime. This framework
comprises closed form expression of the model through tasks and proxies for Transfer
Learning, generalisation and tasks similarity. In this framework, we prove that OGD is robust …
Showing the best results for this search. See all results