Forward Compatible Few-Shot Class-Incremental Learning

Da-Wei Zhou

Nanjing University

Fu-Yun Wang

Nanjing University

Han-Jia Ye

Nanjing University

Liang Ma

Hikvision Research Institute

Shiliang Pu

Hikvision Research Institute

De-Chuan Zhan

Nanjing University


Abstract

Novel classes frequently arise in our dynamically changing world, e.g., new users in the authentication system, and a machine learning model should recognize new classes without forgetting old ones. This scenario becomes more challenging when new class instances are insufficient, which is called few-shot class-incremental learning (FSCIL). Current methods handle incremental learning retrospectively by making the updated model similar to the old one. By contrast, we suggest learning prospectively to prepare for future updates, and propose ForwArd Compatible Training (FACT) for FSCIL. Forward compatibility requires future new classes to be easily incorporated into the current model based on the current stage data, and we seek to realize it by reserving embedding space for future new classes. In detail, we assign virtual prototypes to squeeze the embedding of known classes and reserve for new ones. Besides, we forecast possible new classes and prepare for the updating process. The virtual prototypes allow the model to accept possible updates in the future, which act as proxies scattered among embedding space to build a stronger classifier during inference. FACT efficiently incorporates new classes with forward compatibility and meanwhile resists forgetting of old ones. Extensive experiments on benchmark and large scale datasets validate FACT's state-of-the-art performance.

Framework

Illustration of FACT. Left: making the model growable. Apart from the cross-entropy loss (L1), the model also assigns an instance to a virtual class (L2), which reserves the space for new classes. Right: making the model provident. We first forecast virtual instances by manifold mixup (shown with arrow) and then conduct a symmetric reserving process by assigning it to the virtual class and known class. The training target is a bimodal distribution, which forces the instance to be assigned to different clusters and reserve embedding space.

Selected Results


Video


Da-Wei Zhou, Fu-Yun Wang, Han-Jia Ye, Liang Ma, Shiliang Pu, De-Chuan Zhan
Forward Compatible Few-Shot Class-Incremental Learning
In CVPR, 2022.

[Paper] [Bibtex] [Code]