MetaGCD: Learning to Continually Learn in Generalized Category Discovery

Yanan Wu, Zhixiang Chi, Yang Wang, Songhe Feng; Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2023, pp. 1655-1665

Abstract


In this paper, we consider a real-world scenario where a model that is trained on pre-defined classes continually encounters unlabeled data that contains both known and novel classes. The goal is to continually discover novel classes while maintaining the performance in known classes. We name the setting Continual Generalized Category Discovery (C-GCD). Existing methods for novel class discovery cannot directly handle the C-GCD setting due to some unrealistic assumptions, such as the unlabeled data only containing novel classes. Furthermore, they fail to discover novel classes in a continual fashion. In this work, we lift all these assumptions and propose an approach, called MetaGCD, to learn how to incrementally discover with less forgetting. Our proposed method uses a meta-learning framework and leverages the offline labeled data to simulate the testing incremental learning process. A meta-objective is defined to revolve around two conflicting learning objectives to achieve novel class discovery without forgetting. Furthermore, a soft neighborhood-based contrastive network is proposed to discriminate uncorrelated images while attracting correlated images. We build strong baselines and conduct extensive experiments on three widely used benchmarks to demonstrate the superiority of our method.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Wu_2023_ICCV, author = {Wu, Yanan and Chi, Zhixiang and Wang, Yang and Feng, Songhe}, title = {MetaGCD: Learning to Continually Learn in Generalized Category Discovery}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)}, month = {October}, year = {2023}, pages = {1655-1665} }