[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1007/978-981-97-5492-2_6guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
Article

Meta-pruning: Learning to Prune on Few-Shot Learning

Published: 16 August 2024 Publication History

Abstract

Few-shot learning aims to use a limited amount of data to complete the model’s training, but complex models often face overfitting. To mitigate overfitting risks, we propose a new meta-learning method termed Meta-Pruning, which diverges from traditional pruning methods by treating pruning as a learnable task and training the model to discern and select beneficial network connections for new tasks. We propose to set the corresponding learning rate for each neural network parameter, use the learning rate as a learnable parameter, and adopt a new parameter updating method that reduces reliance on initial learning rate settings and lowers the manual tuning burden. Empirical evaluations on data sets indicate that our proposed algorithm achieves higher classification accuracy with similar computational complexity than state-of-the-art algorithms. We have also conducted learning rate comparison experiments to prove that the initial learning rate no longer affects our method.

References

[1]
Jiang, S., et al.: Explainable text classification via attentive and targeted mixing data augmentation, vol. 2023-August, (Macao, China), pp. 5085–5094 (2023). Attention mechanisms; Augmentation methods; Classifieds; Data augmentation; Data Selection; Loss functions; Model training; Selection framework; Text classification; Training sample;
[2]
Zhong T, Chi Z, Gu L, Wang Y, Yu Y, and Tang J Meta-DMoE: adapting to domain shift by meta-distillation from mixture-of-experts Adv. Neural. Inf. Process. Syst. 2022 35 22243-22257
[3]
Chu, Y., Sun, X., Songhao, J., Xie, T., Wang, Z., Shan, W.: Imbalanced few-shot learning based on meta-transfer learning. In: International Conference on Artificial Neural Networks, pp. 357–369. Springer (2023).
[4]
Chen J, Zhang R, Mao Y, and Xu J ContrastNet: a contrastive learning framework for few-shot text classification Proc. AAAI Conf. Artif. Intell. 2022 36 10492-10500
[5]
Liu J et al. Discrimination-aware network pruning for deep model compression IEEE Trans. Pattern Anal. Mach. Intell. 2021 44 8 4035-4051
[6]
Yu, S., Mazaheri, A., Jannesari, A.: Topology-aware network pruning using multi-stage graph embedding and reinforcement learning. In: International Conference on Machine Learning, pp. 25656–25667, PMLR (2022)
[7]
He, Z., Xie, Z., Zhu, Q., Qin, Z.: Sparse double descent: where network pruning aggravates overfitting. In: International Conference on Machine Learning, pp. 8635–8659, PMLR (2022)
[8]
Geng, B., Yuan, F., Xu, Q., Shen, Y., Xu, R., Yang, M.: Continual learning for task-oriented dialogue system with iterative network pruning, expanding and masking. Ann. Meeting Assoc. Comput. Linguist., 517–523 (2021). abs/2107.08173
[9]
Guo, D., Rush, A.M., Kim, Y.: Parameter-efficient transfer learning with diff pruning. Ann. Meeting Assoc. Comput. Linguist. (2020)
[10]
Nilsback, M.-E., Zisserman, A.: Automated flower classification over a large number of classes. In: 2008 Sixth Indian Conference on Computer Vision, Graphics Image Processing, pp. 722–729. IEEE (2008)
[11]
Vinyals, O., Blundell, C., Lillicrap, T.P., kavukcuoglu, K., Wierstra, D.: Matching networks for one shot learning. Adv. Neural Inf. Process. Syst. 29, 3637–3645 (2016)
[12]
Lake BM, Salakhutdinov R, and Tenenbaum JB Human-level concept learning through probabilistic program induction Science 2015 350 6266 1332-1338
[13]
Bertinetto, L., Henriques, J.F., Torr, P.H.S., Vedaldi, A.: Meta-learning with differentiable closed-form solvers. Int. Conf. Learn. Represent. (2019). abs/1805.08136

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Guide Proceedings
Knowledge Science, Engineering and Management: 17th International Conference, KSEM 2024, Birmingham, UK, August 16–18, 2024, Proceedings, Part I
Aug 2024
461 pages
ISBN:978-981-97-5491-5
DOI:10.1007/978-981-97-5492-2
  • Editors:
  • Cungeng Cao,
  • Huajun Chen,
  • Liang Zhao,
  • Junaid Arshad,
  • Taufiq Asyhari,
  • Yonghao Wang

Publisher

Springer-Verlag

Berlin, Heidelberg

Publication History

Published: 16 August 2024

Author Tags

  1. few-shot learning
  2. meta-learning
  3. pruning
  4. overfitting

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 0
    Total Downloads
  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 11 Dec 2024

Other Metrics

Citations

View Options

View options

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media