Class-incremental Learning using a Sequence of Partial Implicitly Regularized Classifiers
DOI:
https://doi.org/10.32473/flairs.v35i.130549Keywords:
continual-learning, class-incremental learning, catastrophic forgettingAbstract
In class-incremental learning, the objective is to learn a number of classes sequentially without having access to the whole training data. However, due to a problem known as catastrophic forgetting, neural networks suffer substantial performance drop in such settings. The problem is often approached by experience replay, a method that stores a limited number of samples to be replayed in future steps to reduce forgetting of the learned classes. When using a pretrained network as a feature extractor, we show that instead of training a single classifier incrementally, it is better to train a number of specialized classifiers which do not interfere with each other yet can cooperatively predict a single class. Our experiments on CIFAR100 dataset show that the proposed method improves the performance over SOTA by a large margin.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2022 Sobirdzhon Bobiev, Albina Khusainova, Adil Khan, S.M. Ahsan Kazmi
This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.