Abstract
In the few-shot class-incremental learning, new class samples are utilized to learn the characteristics of new classes, while old class exemplars are used to avoid old knowledge forgetting. The limited number of new class samples is more likely to cause overfitting during incremental training. Moreover, mass stored old exemplars mean large storage space consumption. To solve the above difficulties, in this paper we propose a novel flexible few-shot class-incremental framework to make the incremental process efficient and convenient. We enhance the expression ability of extracted features through multistage pre-training. Then, we set up a prototype container to store each class prototype to retain old knowledge. When new classes flow in, we calculate the new class prototypes and update the prototype container. Finally, we get the prediction result through similarity weighting. The entire framework only need to train the base class classifier and does not require further training during the incremental process. It avoids the overfitting of novel classes and saves time for further training. Besides, storing prototypes can save more storage space than original image data. Overall, the entire framework has the advantage of flexibility. We conduct extensive experiments on three standard few-shot class-incremental datasets and achieve state-of-the-art results. Especially, to verify the flexibility of the framework, we discuss the special federated few-shot class-incremental scenarios in addition. No further training and less storage consumption provide the possibility for applications in more complex scenarios.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
The datasets generated during and/or analyzed during the current study are available in the [The CIFAR100 dataset] repository with [http://www.cs.toronto.edu/~kriz/cifar.html], [Caltech-UCSD Birds-200-2011 (CUB-200-2011)] repository with [https://www.vision.caltech.edu/datasets/cub_200_2011/] and [miniImageNet] repository with [https://www.kaggle.com/datasets/arjunashok33/miniimagenet].
References
McCormack J, Lomas A (2021) Deep learning of individual aesthetics. Neural Comput Appl 33(1):3–17
Chen T, Frankle J, Chang S, Liu S, Zhang Y, Carbin M, Wang Z (2021) The lottery tickets hypothesis for supervised and self-supervised pre-training in computer vision models. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 16306–16316
Zhang F, Xu X, Nauata N, Furukawa Y (2021) Structured outdoor architecture reconstruction by exploration and classification. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 12427–12435
Joseph K, Khan S, Khan FS, Balasubramanian VN (2021) Towards open world object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 5830–5840
Ye J, Yang X, Kang S, He Y, Zhang W, Huang L, Jiang M, Zhang W, Shi Y, Xia M et al (2021) A robust MTMC tracking system for AI-city challenge 2021. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4044–4053
Wang Z, Fang Z, Li D, Yang H, Du W (2021) Semantic supplementary network with prior information for multi-label image classification. IEEE Trans Circuits Syst Video Technol 32(4):1848–1859
Zhu Y, Wang Z, Zha H, Gao D (2017) Boundary-eliminated pseudoinverse linear discriminant for imbalanced problems. IEEE Trans Neural Netw Learn Syst 29(6):2581–2594
Izzuddin TA, Safri NM, Othman MA (2021) Mental imagery classification using one-dimensional convolutional neural network for target selection in single-channel BCI-controlled mobile robot. Neural Comput Appl 33(11):6233–6246
Ruping S (2001) Incremental learning with support vector machines. In: Proceedings 2001 IEEE international conference on data mining, IEEE, pp 641–642
Wu Y, Chen Y, Wang L, Ye Y, Liu Z, Guo Y, Fu Y (2019) Large scale incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 374–382
Castro FM, Marín-Jiménez MJ, Guil N, Schmid C, Alahari K (2018) End-to-end incremental learning. In: Proceedings of the European conference on computer vision, pp 233–248
Polikar R, Upda L, Upda SS, Honavar V (2001) Learn++: an incremental learning algorithm for supervised neural networks. IEEE Trans Syst Man Cybern Part C Appl Rev 31(4):497–508
Wang Z, Cao C, Zhu Y (2020) Entropy and confidence-based undersampling boosting random forests for imbalanced problems. IEEE Trans Neural Netw Learn Syst 31(12):5178–5191
Michieli U, Zanuttigh P (2021) Knowledge distillation for incremental learning in semantic segmentation. Comput Vis Image Underst 205:103167
Javed K, Shafait F (2018) Revisiting distillation and incremental classifier learning. In: Asian conference on computer vision, Springer, pp 3–17
Chen L, Yu C, Chen L (2019) A new knowledge distillation for incremental object detection. In: 2019 International joint conference on neural networks (IJCNN), IEEE, pp 1–7
Xiang Y, Miao Y, Chen J, Xuan Q (2020) Efficient incremental learning using dynamic correction vector. IEEE Access 8:23090–23099
Valipour S, Perez C, Jagersand M (2017) Incremental learning for robot perception through HRI. In: 2017 IEEE/RSJ international conference on intelligent robots and systems, IEEE, pp 2772–2777
Han S, Meng Z, Khan A-S, Tong Y (2016) Incremental boosting convolutional neural network for facial action unit recognition. In: Advances in neural information processing systems, vol 29, pp 109–117
Perez E, Kiela D, Cho K (2021) True few-shot learning with language models. In: Advances in neural information processing systems, vol 34, pp 11054–11070
Sun X, Wang B, Wang Z, Li H, Li H, Fu K (2021) Research progress on few-shot learning for remote sensing image interpretation. IEEE J Sel Top Appl Earth Obs Remote Sens 14:2387–2402
Ma J, Fong SH, Luo Y, Bakkenist CJ, Shen JP, Mourragui S, Wessels LF, Hafner M, Sharan R, Peng J et al (2021) Few-shot learning creates predictive models of drug response that translate from high-throughput screens to individual patients. Nat Cancer 2(2):233–244
Tao X, Hong X, Chang X, Dong S, Wei X, Gong Y (2020) Few-shot class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 12183–12192
Cheraghian A, Rahman S, Fang P, Roy SK, Petersson L, Harandi M (2021) Semantic-aware knowledge distillation for few-shot class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2534–2543
Li Y, Yang J (2021) Meta-learning baselines and database for few-shot classification in agriculture. Comput Electron Agric 182:106055
Boudiaf M, Kervadec H, Masud ZI, Piantanida P, Ben Ayed I, Dolz J (2021) Few-shot segmentation without meta-learning: a good transductive inference is all you need? In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 13979–13988
Chen Y, Liu Z, Xu H, Darrell T, Wang X (2021) Meta-baseline: exploring simple meta-learning for few-shot learning. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 9062–9071
Koch G, Zemel R, Salakhutdinov R et al (2015) Siamese neural networks for one-shot image recognition. In: ICML deep learning workshop, Lille, vol 2
Snell J, Swersky K, Zemel R (2017) Prototypical networks for few-shot learning. In: Proceedings of the 31st international conference on neural information processing systems, pp 4080–4090
Sung F, Yang Y, Zhang L, Xiang T, Torr PH, Hospedales TM (2018) Learning to compare: relation network for few-shot learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1199–1208
Vinyals O, Blundell C, Lillicrap T, Wierstra D et al (2016) Matching networks for one shot learning. Adv Neural Inf Process Syst 29:3630–3638
Zhu K, Cao Y, Zhai W, Cheng J, Zha Z-J (2021) Self-promoted prototype refinement for few-shot class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6801–6810
Zhang C, Song N, Lin G, Zheng Y, Pan P, Xu Y (2021) Few-shot incremental learning with continually evolved classifiers. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 12455–12464
Chi Z, Gu L, Liu H, Wang Y, Yu Y, Tang J (2022) Metafscil: a meta-learning approach for few-shot class incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 14166–14175
Zhou D-W, Wang F-Y, Ye H-J, Ma L, Pu S, Zhan D-C (2022) Forward compatible few-shot class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9046–9056
Huang H, Wu Z, Li W, Huo J, Gao Y (2021) Local descriptor-based multi-prototype network for few-shot learning. Pattern Recognit 116:107935
Ji Z, Chai X, Yu Y, Zhang Z (2021) Reweighting and information-guidance networks for few-shot learning. Neurocomputing 423:13–23
Deuschel J, Firmbach D, Geppert CI, Eckstein M, Hartmann A, Bruns V, Kuritcyn P, Dexl J, Hartmann D, Perrin D et al (2021) Multi-prototype few-shot learning in histopathology. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 620–628
Zimmermann RS, Sharma Y, Schneider S, Bethge M, Brendel W (2021) Contrastive learning inverts the data generating process. In: International conference on machine learning, PMLR, pp 12979–12990
Zhong Z, Fini E, Roy S, Luo Z, Ricci E, Sebe N (2021) Neighborhood contrastive learning for novel class discovery. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10867–10875
Diba A, Sharma V, Safdari R, Lotfi D, Sarfraz S, Stiefelhagen R, Van Gool L (2021) Vi2clr: video and image for visual contrastive learning of representation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 1502–1512
Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M et al (2015) Imagenet large scale visual recognition challenge. Int J Comput Vis 115(3):211–252
Wah C, Branson S, Welinder P, Perona P, Belongie S (2011) The caltech-ucsd birds-200-2011 dataset. Technical Report CNS-TR-2011-001, California Institute of Technology
Zhang C, Cai Y, Lin G, Shen C (2020) Deepemd: few-shot image classification with differentiable earth mover’s distance and structured classifiers. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 12203–12213
Rebuffi S-A, Kolesnikov A, Sperl G, Lampert C H (2017) ICARL: incremental classifier and representation learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2001–2010
Hou S, Pan X, Loy CC, Wang Z, Lin D, (2019) Learning a unified classifier incrementally via rebalancing. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 831–839
Liu J, Song L, Qin Y, Prototype rectification for few-shot learning (2020). In: Computer vision–ECCV 2020: 16th European conference, Glasgow, UK, 23–28 Aug 2020, proceedings, part I 16, Springer, pp 741–756
Chen Y, Wang X, Liu Z, Xu H, Darrell T (2020) A new meta-baseline for few-shot learning. arXiv:2003.04390
Qi H, Brown M, Lowe DG (2018) Low-shot learning with imprinted weights. In: Proceedings of the IEEE conference on computer Cision and pattern recognition, pp 5822–5830
Acknowledgements
This work is supported by Shanghai Science and Technology Program “Federated based cross-domain and cross-task incremental learning” under Grant No. 21511100800, Shanghai Science and Technology Program ”Distributed and generative few-shot algorithm and theory research” under Grant No. 20511100600, Natural Science Foundation of China under Grant No. 62076094 and No. 62276098. Chinese Defense Program of Science and Technology under Grant No.2021-JCJQ-JJ-0041, China Aerospace Science and Technology Corporation Industry-University-Research Cooperation Foundation of the Eighth Research Institute under Grant No. SAST2021-007.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Xu, X., Wang, Z., Fu, Z. et al. Flexible few-shot class-incremental learning with prototype container. Neural Comput & Applic 35, 10875–10889 (2023). https://doi.org/10.1007/s00521-023-08272-y
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-023-08272-y