On the Soft-Subnetwork for Few-Shot Class Incremental Learning

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 437
  • Download : 0
Inspired by Regularized Lottery Ticket Hypothesis, which states that competitive smooth (non-binary) subnetworks exist within a dense network, we propose a few-shot class-incremental learning method referred to as Soft-SubNetworks (SoftNet). Our objective is to learn a sequence of sessions incrementally, where each session only includes a few training instances per class while preserving the knowledge of the previously learned ones. SoftNet jointly learns the model weights and adaptive non-binary soft masks at a base training session in which each mask consists of the major and minor subnetwork; the former aims to minimize catastrophic forgetting during training, and the latter aims to avoid overfitting to a few samples in each new training session. We provide comprehensive empirical validations demonstrating that our SoftNet effectively tackles the few-shot incremental learning problem by surpassing the performance of state-of-the-art baselines over benchmark datasets.
Publisher
International Conference on Learning Representations
Issue Date
2023-05-01
Language
English
Citation

International Conference on Learning Representations (ICLR) 2023

URI
http://hdl.handle.net/10203/307225
Appears in Collection
AI-Conference Papers(학술대회논문)EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0