[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Class-incremental learning via prototype similarity replay and similarity-adjusted regularization

Published: 30 July 2024 Publication History

Abstract

The task of incremental learning is to enable machine learning models to continuously learn and adapt to new tasks and data in changing environments while maintaining knowledge of prior tasks. Recently, researchers have proposed a variety of incremental learning methods. Some methods rely on data storage or complex generative models to perform satisfactorily. However, existing incremental learning approaches typically focus on mitigating catastrophic forgetting, with less emphasis on effectively applying old knowledge to facilitate learning new tasks. In this paper, we propose a non-exemplar-based incremental learning approach called Class-Incremental Learning via Prototype Similarity Replay and Similarity-adjusted Regularization (PSSR) to tackle catastrophic forgetting in incremental learning. The essence of PSSR is leveraging prior knowledge of prior tasks to facilitate the acquisition of new tasks. PSSR memorizes a prototype for each old class, representing the class, and learns the new classes based on the similarity between the prototypes and the new class samples during the learning process. The feature space distribution is modified by the old class prototypes to enhance the model’s learning of the new classes. Extensive experiments on three benchmark datasets demonstrate the superior incremental performance of PSSR, with classification accuracy improvements of 2.73%, 3.37%, and 4.21% over state-of-the-art methods. Code available at https://github.com/FutureIAI/PSSR.

References

[1]
Turkoglu B, Uymaz SA, Kaya E (2023) Chaos theory in metaheuristics. In: Comprehensive metaheuristics (pp 1–20). Academic, Cambridge
[2]
Koçer HG, Türkoğlu B, and Uymaz SA Chaotic golden ratio guided local search for big data optimization Eng Sci Technol Int J 2023 41 101388
[3]
Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition. IEEE, pp 248–255
[4]
Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25:1097–1105
[5]
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770–778)
[6]
Ali A, Zhu Y, and Zakarya M Exploiting dynamic spatio-temporal graph convolutional neural networks for citywide traffic flows prediction Neural Netw 2022 145 233-247
[7]
Zhou DW, Yang Y, and Zhan DC Learning to classify with incremental new class IEEE Trans Neural Netw Learn Syst 2021 33 6 2429-2443
[8]
Zhou DW, Ye HJ, Zhan DC (2021) Learning placeholders for open-set recognition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp 4401–4410)
[9]
Wang L, Zhang X, Su H, Zhu J (2024) A comprehensive survey of continual learning: theory, method and application. IEEE Trans Pattern Anal Mach Intell.
[10]
Zhou DW, Wang QW, Qi ZH, Ye HJ, Zhan DC, Liu Z (2023) Deep class-incremental learning: a survey. arxiv preprint arxiv:2302.03648
[11]
Bang J, Kim H, Yoo Y, Ha JW, Choi J (2021) Rainbow memory: Continual learning with a memory of diverse samples. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 8218–8227)
[12]
Sun W, Li Q, Zhang J, Wang D, Wang W, and Geng YA Exemplar-free class incremental learning via discriminative and comparable parallel one-class classifiers Pattern Recogn 2023 140
[13]
Petit G, Popescu A, Schindler H, Picard D, Delezoide B (2023) Fetril: Feature translation for exemplar-free class-incremental learning. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision (pp 3911–3920)
[14]
Wang S, Li X, Sun J, Xu Z (2021) Training networks in null space of feature covariance for continual learning. In: Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition (pp 184–193)
[15]
Zhou DW, Wang QW, Ye HJ, Zhan DC (2022) A model or 603 exemplars: towards memory-efficient class-incremental learning. arxiv preprint arxiv:2205.13218
[16]
Wang Z, Zhang Z, Lee CY, Zhang H, Sun R, Ren X, Su G, Perot V, Dy J, Pfister T (2022) Learning to prompt for continual learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp 139–149)
[17]
Xie J, Yan S, He X (2022) General incremental learning with domain-aware categorical representations. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp 14351–14360)
[18]
Lu Y, Wang M, Deng W (2022) Augmented geometric distillation for data-free incremental person reid. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp 7329–7338)
[19]
Gao Q, Zhao C, Ghanem B, Zhang J (2022) R-dfcil: Relation-guided representation learning for data-free class incremental learning. In: European Conference on Computer Vision (pp 423–439). Cham: Springer Nature Switzerland
[20]
Zhou DW, Wang FY, Ye HJ, Ma L, Pu S, Zhan DC (2022) Forward compatible few-shot class-incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp 9046–9056)
[21]
Díaz-Rodríguez N, Lomonaco V, Filliat D, Maltoni D (2018) Don’t forget, there is more than forgetting: new metrics for Continual Learning. arXiv preprint arXiv:1810.13166
[22]
Lopez-Paz D, Ranzato MA (2017) Gradient episodic memory for continual learning. Adv Neural Inf Process Syst 30:6470–6479
[23]
Lei CH, Chen YH, Peng WH, Chiu WC (2020) Class-incremental learning with rectified feature-graph preservation. In: Proceedings of the Asian Conference on Computer Vision
[24]
Shi Y, Zhou K, Liang J, Jiang Z, Feng J, Torr PH, … Tan VY (2022) Mimicking the oracle: an initial phase decorrelation approach for class incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 16722–16731
[25]
Hersche M, Karunaratne G, Cherubini G, Benini L, Sebastian A, Rahimi A (2022) Constrained few-shot class-incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 9057–9067
[26]
Zhu K, Zhai W, Cao Y, Luo J, Zha ZJ (2022) Self-sustaining representation expansion for non-exemplar class-incremental learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 9296–9305
[27]
Smith J, Hsu YC, Balloch J, Shen Y, Jin H, Kira Z (2021) Always be dreaming: a new approach for data-free class-incremental learning. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp 9374–9384
[28]
Jin H, Kim E (2022) Helpful or harmful: inter-task association in continual learning. In: European Conference on Computer Vision. Springer Nature Switzerland, Cham, pp 519–535
[29]
Zhu F, Cheng Z, Zhang XY, and Liu CL Class-incremental learning via dual augmentation Adv Neural Inf Process Syst 2021 34 14306-14318
[30]
Zhu F, Zhang XY, Wang C, Yin F, Liu CL (2021) Prototype augmentation and self-supervision for incremental learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp 5871–5880)
[31]
Shi Y, Shi D, Qiao Z, Wang Z, Zhang Y, Yang S, and Qiu C Multi-granularity knowledge distillation and prototype consistency regularization for class-incremental learning Neural Netw 2023 164 617-630
[32]
Zhang W and Gu X Few shot class incremental learning via efficient prototype replay and calibration Entropy 2023 25 5
[33]
Mao K, Luo Y, Ren Y, and Wang R Prototype representation expansion in incremental learning Neural Process Lett 2023 55 6 8401-8417
[34]
Asadi N, Davari M, Mudur S, Aljundi R, Belilovsky E (2023) Prototype-sample relation distillation: towards replay-free continual learning. In: International Conference on Machine Learning. PMLR, pp 1093–1106
[35]
Shen M, Chen D, Hu S, and Xu G Class incremental learning of remote sensing images based on class similarity distillation PeerJ Comput Sci 2023 9 e1583
[36]
Bansal G, Nushi B, Kamar E, Weld DS, Lasecki WS, Horvitz E (2019) Updates in human-ai teams: Understanding and addressing the performance/compatibility tradeoff. In: Proceedings of the AAAI Conference on Artificial Intelligence (vol 33, No 01, pp 2429–2437)
[37]
Srivastava M, Nushi B, Kamar E, Shah S, Horvitz E (2020) An empirical analysis of backward compatibility in machine learning systems. In: Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. pp 3272–3280
[38]
Hou S, Pan X, Loy CC, Wang Z, Lin D (2019) Learning a unified classifier incrementally via rebalancing. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp 831–839)
[39]
Richie R and Bhatia S Similarity judgment within and across categories: a comprehensive model comparison Cogn Sci 2021 45 8
[40]
Krizhevsky A, Hinton G (2009) Learning multiple layers of features from tiny images
[41]
Yao L, Miller J (2015) Tiny imagenet classification with convolutional neural networks. CS 231 N(5):82
[42]
Cohen G, Afshar S, Tapson J, Van Schaik A (2017) EMNIST: Extending MNIST to handwritten letters. In: 2017 international joint conference on neural networks (IJCNN) (pp 2921–2926). IEEE
[43]
Wah C, Branson S, Welinder P, Perona P, Belongie S (2011) The caltech-ucsd birds-200-2011 dataset
[44]
Xiao H, Rasul K, Vollgraf R (2017) Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747
[45]
Buzzega P, Boschini M, Porrello A, Abati D, and Calderara S Dark experience for general continual learning: a strong, simple baseline Adv Neural Inf Process Syst 2020 33 15920-15930
[46]
Kingma DP, Ba J (2015) Adam: a method for stochastic optimization. In: 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015, Conference track proceedings
[47]
Kirkpatrick J, Pascanu R, Rabinowitz N, Veness J, Desjardins G, Rusu AA, Kirkpatrick J, Pascanu R, Rabinowitz N, Veness J, Desjardins G, Rusu AA, Milan K, Quan J, Ramalho T, Grabska-Barwinska A, and Hadsell R Overcoming catastrophic forgetting in neural networks Proceedings of the National Academy of Sciences 2017 114 13 3521-3526
[48]
Li Z and Hoiem D Learning without forgetting IEEE Trans Pattern Anal Mach Intell 2017 40 12 2935-2947
[49]
Liu Y, Parisot S, Slabaugh G, Jia X, Leonardis A, Tuytelaars T (2020) More classifiers, less forgetting: a generic multi-classifier paradigm for incremental learning. In: Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXVI 16. Springer International Publishing, pp 699–716
[50]
Rebuffi SA, Kolesnikov A, Sperl G, Lampert CH (2017) icarl: Incremental classifier and representation learning. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition. pp 2001–2010
[51]
Castro FM, Marín-Jiménez MJ, Guil N, Schmid C, Alahari K (2018) End-to-end incremental learning. In: Proceedings of the European Conference on Computer Vision (ECCV). pp 233–248

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Applied Intelligence
Applied Intelligence  Volume 54, Issue 20
Oct 2024
715 pages

Publisher

Kluwer Academic Publishers

United States

Publication History

Published: 30 July 2024
Accepted: 15 July 2024

Author Tags

  1. Class incremental learning
  2. Transfer forward
  3. Bayes’ theorem
  4. Similar feature

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 0
    Total Downloads
  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 01 Jan 2025

Other Metrics

Citations

View Options

View options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media