[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ Skip to main content
Log in

Few-shot learning with hierarchical pooling induction network

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Learning to recognize new concepts from few-shot examples is a long-standing challenge in modern computer vision. Metric based few-shot learning is a prevalent way towards this goal where new query instances are classified to the support classes by comparing the query instance with each class prototype. To this end, a representative and discriminative class prototype should necessarily be induced from several support example features. In this work, we propose a simple but effective hierarchical pooling induction method to learn such generalized class-level representations, by concatenating the max pooling and mean pooling operations. The proposed induction method could form a representative prototype for given few-shot samples, enhancing both the discrimination of the intermediate features and the final classification performance. The benchmark miniImageNet dataset and some other practical Remote Sensing Image Scene Classification (RESISC) datasets are employed for generalization investigation and it shows that the proposed induction module could improve the performance of state-of-the-art method and outperforms other alternative induction methods. Qualitative visualization and quantitative analysis are also provided to demonstrate the effectiveness and robustness of the proposed method.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
£29.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (United Kingdom)

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

References

  1. Battaglia P, Hamrick J B, Bapst V et al (2018) Relational inductive biases, deep learning, and graph networks[J]. arXiv:Learning

  2. Cai Q, Pan Y, Yao T et al (2018) Memory matching networks for one-shot image recognition[C]. Comput Vis Pattern Recogn:4080–4088

  3. Cheng G, Han J, Lu X et al (2017) Remote sensing image scene classification: benchmark and state of the art[J]. arXiv: Comput Vis Pattern Recogn 105 (10):1865–1883

    Google Scholar 

  4. Finn C, Abbeel P, Levine S et al (2017) Model-agnostic meta-learning for fast adaptation of deep networks[C]. Int Conf Mach Learn:1126–1135

  5. Gao H, Shou Z, Zareian A et al (2018) Low-shot Learning via Covariance-Preserving Adversarial Augmentation Networks.[J] arXiv: Machine Learning

  6. Garcia V, Bruna J. (2017) Few-Shot Learning with Graph Neural Networks.[J] arXiv: Machine Learning

  7. Geng R, Li B, Li Y et al (2019) Few-Shot Text Classification with Induction Network.[J]. arXiv:Computation and Language

  8. Grant E, Finn C, Levine S et al (2018) Recasting Gradient-Based Meta-Learning as Hierarchical Bayes[C]. International Conference on Learning Representations

  9. Hariharan B, Girshick R B (2017) Low-shot visual recognition by shrinking and hallucinating features[C]. Int Conf Comput Vis:3037–3046

  10. He K, Zhang X, Ren S et al (2016) Deep residual learning for image recognition[C]. Comput Vis Pattern Recogn:770–778

  11. Iyyer M, Manjunatha V, Boydgraber J L et al (2015) Deep unordered composition rivals syntactic methods for text classification[C]. Int Joint Conf Natural Lang Process:1681–1691

  12. Joulin A, Grave E, Bojanowski P et al (2017) Bag of tricks for efficient text classification[C]. Conf Eur Chapter Assoc Comput Linguist:427–431

  13. Kaiser L, Nachum O, Roy A et al (2017) Learning to Remember Rare Events[C]. International Conference on Learning Representations

  14. Koch G, Zemel R, Salakhutdinov R. (2015) Siamese neural networks for one-shot image recognition[C]. ICML Deep Learning workshop

  15. Krizhevsky A, Sutskever I, Hinton G E et al (2012) ImageNet Classification with Deep Convolutional Neural Networks[C]. Neural Inf Process Syst 141 (5):1097–1105

    Google Scholar 

  16. Lake B M, Salakhutdinov R, Tenenbaum J B et al (2015) Human-level concept learning through probabilistic program induction[J]. Science 350 (6266):1332–1338

    Article  MathSciNet  Google Scholar 

  17. Lecun Y, Bengio Y, Hinton G E et al (2015) Deep learning[J]. Nature 521(7553):436–444

    Article  Google Scholar 

  18. Lee Y (2018) Choi S. Gradient-based meta-learning with learned layerwise metric and subspace[C]. Int Conf Mach Learn:2927–2936

  19. Li Z, Zhou F, Chen F et al (2017) Meta-SGD: Learning to Learn Quickly for Few Shot Learning[J]. arXiv: Learning

  20. Liu Y, Lee J, Park M et al (2019) Learning to propagate labels: transductive propagation network for few-shot learning[C]. International Conference on Learning Representations

  21. Luca B, Joao F H, Jack V et al (2016) Learning feed-forward one-shot learners[C]. Neural Information Processing Systems

  22. Munkhdalai T, Yu H. (2017) Meta Networks.[J] arXiv: Machine Learning

  23. Ravi S, Larochelle H. (2017) Optimization as a Model for Few-Shot Learning[C]. International Conference on Learning Representations

  24. Rusu A A, Rao D, Sygnowski J et al (2019) Meta-Learning with Latent Embedding Optimization[C]. International Conference on Learning Representations

  25. Sabour S, Frosst N, Hinton G E et al (2017) Dynamic routing between capsules[C]. Neural Inf Process Syste:3856–3866

  26. Salvaris M, Dean D, Tok W H et al (2018) Generative Adversarial Networks[J]. arXiv: Machine Learning, pp 187–208

  27. Santoro A, Bartunov S, Botvinick M M et al (2016) One-shot Learning with Memory-Augmented Neural Networks[J]. arXiv: Learning

  28. Schwartz E, Karlinsky L, Shtok J et al (2018) Delta-encoder: an effective sample synthesis method for few-shot object recognition.[J] arXiv: Comput Vis Pattern Recogn

  29. Shen D, Wang G, Wang W et al (2018) Baseline needs more love: on simple word-embedding-based models and associated pooling mechanisms[C]. Meeting Assoc Comput Linguist:440–450

  30. Snell J, Swersky K, Zemel RS et al (2017) Prototypical networks for few-shot learning[C]. Neural Inf Process Syst:4077–4087

  31. Sung W Y, Jun S, Jaekyun M. (2019) TapNet: neural network augmented with task-adaptive projection for few-shot learning[C]. Int Conf Mach Learn:7115–7123

  32. Sung F, Yang Y, Zhang L et al (2018) Learning to Compare: relation network for few-shot learning[C]. Comput Vis Pattern Recogn:1199–1208

  33. The Google image dataset of SIRI-WHU, http://www.lmars.whu.edu.cn/prof_web/zhongyanfei/Num/Google.html, Accessed 27 Sept 2018

  34. Thomas E, Benedikt S, Jan H et al (2020) Meta-learning of neural architectures for few-shot learning[C]. Comput Vis Pattern Recogn, 12362–12372

  35. Tyler R S, Karl R, Michael C M (2018) Adapted deep embeddings: a synthesis of methods for k-shot inductive transfer learning. Neural Information Processing Systems

  36. Vinyals O, Blundell C, Lillicrap T P et al (2016) Matching networks for one shot learning[C]. Neural Inf Process Syst:3637–3645

  37. Wang Y, Girshick R B, Hebert M et al (2018) Low-shot learning from imaginary data[C]. Comput Vis Pattern Recogn:7278–7286

  38. Xia G, Hu J, Hu F et al (2017) AID: A benchmark data set for performance evaluation of aerial scene classification[J]. IEEE Trans Geosci Remote Sens 55(7):3965–3981

    Article  Google Scholar 

  39. Yan S, Zhang S, He X. (2019) A Dual Attention Network with Semantic Embedding for Few-Shot Learning[C] AAAI

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jian Huang.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix A:: Normal weighted induction

Appendix A:: Normal weighted induction

Definition:

For a set of discrete samples (x1, x2,⋯, xn), we first get the mean and variance of the samples, denoted as μ and σ, and then define a normal distribution f \(\sim \) N(μ, σ). The weight for each xi is defined as its probability density wi = f (xi) and their normalization values are

$$ {\bar {w_{i}}} = \frac{{{w_{i}}}}{{\sum\limits_{i} {{w_{i}}} }} $$
(4)

Finally, we define the normal weighted induction as:

$$ {p^{Norm}} = \sum\limits_{i} {{\bar {w_{i}}}{x_{i}}} $$
(5)

For a multi-dimensional vector or tensor, we perform the bitwise operations over each dimension, resulting in a vector or tensor of the same dimension.

To qualitatively demonstrate the normal weighted induction, we present the visualization results compared with the unbiased average pooling. Figure 7 shows the 2D samples as well as their induction prototypes, where the red stars indicate the samples and the two induction prototypes are indicated by the cross and circle, respectively.

Fig. 7
figure 7

Visualization of the normal weighted induction. Examples (Indicated by the red stars) are sampled from a normal distribution N((40, 60), (5, 8)), together with 3 apart outliers. The normal weighted induction prototype is closer to the real sampling center than the mean value under some disturbances

We can see that although the unbiased estimation it is, the mean value deviates far from the real sampling center (40, 60), due to the disturbance of some singular points. However, with the normal weighted induction, the prototype gets closer to the sampling center, alleviating the influence of the outliers. Conclusions could be made that the normal weighted induction will be more helpful in the cases that there are some obvious outliers.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Pan, C., Huang, J., Gong, J. et al. Few-shot learning with hierarchical pooling induction network. Multimed Tools Appl 81, 32937–32952 (2022). https://doi.org/10.1007/s11042-022-11999-w

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-022-11999-w

Keywords

Navigation