[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3366423.3380153acmconferencesArticle/Chapter ViewAbstractPublication PagesthewebconfConference Proceedingsconference-collections
research-article

Few-Sample and Adversarial Representation Learning for Continual Stream Mining

Published: 20 April 2020 Publication History

Abstract

Deep Neural Networks (DNNs) have primarily been demonstrated to be useful for closed-world classification problems where the number of categories is fixed. However, DNNs notoriously fail when tasked with label prediction in a non-stationary data stream scenario, which has the continuous emergence of the unknown or novel class (categories not in the training set). For example, new topics continually emerge in social media or e-commerce. To solve this challenge, a DNN should not only be able to detect the novel class effectively but also incrementally learn new concepts from limited samples over time. Literature that addresses both problems simultaneously is limited. In this paper, we focus on improving the generalization of the model on the novel classes, and making the model continually learn from only a few samples from the novel categories. Different from existing approaches that rely on abundant labeled instances to re-train/update the model, we propose a new approach based on Few Sample and Adversarial Representation Learning (FSAR). The key novelty is that we introduce the adversarial confusion term into both the representation learning and few-sample learning process, which reduces the over-confidence of the model on the seen classes, further enhance the generalization of the model to detect and learn new categories with only a few samples. We train the FSAR operated in two stages: first, FSAR learns an intra-class compacted and inter-class separated feature embedding to detect the novel classes; next, we collect a few labeled samples belong to the new categories, utilize episode-training to exploit the intrinsic features for few-sample learning. We evaluated FSAR on different datasets, using extensive experimental results from various simulated stream benchmarks to show that FSAR effectively outperforms current state-of-the-art approaches.

References

[1]
Lada A Adamic, Jun Zhang, Eytan Bakshy, and Mark S Ackerman. 2008. Knowledge sharing and yahoo answers: everyone knows something. In Proceedings of the 17th international conference on World Wide Web. ACM, 665–674.
[2]
Hongjoon Ahn, Sungmin Cha, Donggyu Lee, and Taesup Moon. 2019. Uncertainty-based continual learning with adaptive regularization. In NeurIPS. 4394–4404.
[3]
Abhijit Bendale and Terrance E Boult. 2016. Towards open set deep networks. In CVPR. 1563–1572.
[4]
Konstantinos Bousmalis, Nathan Silberman, David Dohan, Dumitru Erhan, and Dilip Krishnan. 2017. Unsupervised pixel-level domain adaptation with generative adversarial networks. In CVPR. 3722–3731.
[5]
Swarup Chandra, Ahsanul Haque, Hemeng Tao, Jie Liu, Latifur Khan, and Charu Aggarwal. 2018. Ensemble direct density ratio estimation for multistream classification. In ICDE. 1364–1367.
[6]
Wei-Yu Chen, Yen-Cheng Liu, Zsolt Kira, Yu-Chiang Frank Wang, and Jia-Bin Huang. 2019. A closer look at few-shot classification. ICLR (2019).
[7]
De Cheng, Yihong Gong, Sanping Zhou, Jinjun Wang, and Nanning Zheng. 2016. Person re-identification by multi-channel parts-based cnn with improved triplet loss function. In CVPR. 1335–1344.
[8]
Terrance DeVries and Graham W Taylor. 2018. Learning Confidence for Out-of-Distribution Detection in Neural Networks. arXiv preprint arXiv:1802.04865(2018).
[9]
Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In ICML. JMLR. org, 1126–1135.
[10]
Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. In NeurIPS. 2672–2680.
[11]
Raia Hadsell, Sumit Chopra, and Yann LeCun. 2006. Dimensionality reduction by learning an invariant mapping. In CVPR. IEEE, 1735–1742.
[12]
Ahsanul Haque, Latifur Khan, and Michael Baron. 2016. SAND: Semi-Supervised Adaptive Novel Class Detection and Classification over Data Stream. In AAAI. 1652–1658.
[13]
A. Haque, L. Khan, M. Baron, B. Thuraisingham, and C. Aggarwal. 2016. Efficient handling of concept drift and concept evolution over Stream Data. In ICDE. 481–492. https://doi.org/10.1109/ICDE.2016.7498264
[14]
Ahsanul Haque, Hemeng Tao, Swarup Chandra, Jie Liu, and Latifur Khan. 2018. A framework for multistream regression with direct density ratio estimation. In AAAI.
[15]
Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. 2017. Mask r-cnn. In ICCV. IEEE, 2980–2988.
[16]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In CVPR. 770–778.
[17]
Dan Hendrycks and Kevin Gimpel. 2016. A baseline for detecting misclassified and out-of-distribution examples in neural networks. ICLR (2016).
[18]
Geoffrey Hinton, Li Deng, Dong Yu, George E Dahl, Abdel-rahman Mohamed, Navdeep Jaitly, Andrew Senior, Vincent Vanhoucke, Patrick Nguyen, Tara N Sainath, 2012. Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups. IEEE Signal processing magazine 29, 6 (2012), 82–97.
[19]
Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531(2015).
[20]
Cheng-Kang Hsieh, Longqi Yang, Yin Cui, Tsung-Yi Lin, Serge Belongie, and Deborah Estrin. 2017. Collaborative metric learning. In Proceedings of the 26th international conference on world wide web. International World Wide Web Conferences Steering Committee, 193–201.
[21]
Jyh-Jing Hwang, Tsung-Wei Ke, Jianbo Shi, and Stella X Yu. 2019. Adversarial Structure Matching for Structured Prediction Tasks. In CVPR. 4056–4065.
[22]
Alex Krizhevsky and Geoffrey Hinton. 2009. Learning multiple layers of features from tiny images. Technical Report. Citeseer.
[23]
Himabindu Lakkaraju, Ece Kamar, Rich Caruana, and Eric Horvitz. 2017. Identifying unknown unknowns in the open world: Representations and policies for guided exploration. In AAAI.
[24]
Shiyu Liang, Yixuan Li, and R Srikant. 2017. Enhancing the reliability of out-of-distribution image detection in neural networks. in ICLR (2017).
[25]
Laurens van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE. Journal of machine learning research 9, Nov (2008), 2579–2605.
[26]
Mohammad Masud, Jing Gao, Latifur Khan, Jiawei Han, and Bhavani M. Thuraisingham. 2011. Classification and Novel Class Detection in Concept-Drifting Data Streams Under Time Constraints. IEEE Trans. on Knowl. and Data Eng. 23, 6 (June 2011), 859–874.
[27]
Nikhil Mishra, Mostafa Rohaninejad, Xi Chen, and Pieter Abbeel. 2018. A simple neural attentive meta-learner. ICML (2018).
[28]
Yair Movshovitz-Attias, Alexander Toshev, Thomas K Leung, Sergey Ioffe, and Saurabh Singh. 2017. No fuss distance metric learning using proxies. In ICCV. 360–368.
[29]
Xin Mu, Feida Zhu, Juan Du, Ee-Peng Lim, and Zhi-Hua Zhou. 2017. Streaming Classification with Emerging New Class by Class Matrix Sketching. In AAAI. 2373–2379.
[30]
Tsendsuren Munkhdalai and Hong Yu. [n.d.]. Meta networks.
[31]
Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y Ng. 2011. Reading digits in natural images with unsupervised feature learning. In NeurIPS workshop on deep learning and unsupervised feature learning, Vol. 2011. 5.
[32]
Anh Nguyen, Jason Yosinski, and Jeff Clune. 2015. Deep neural networks are easily fooled: High confidence predictions for unrecognizable images. In CVPR. 427–436.
[33]
Hyun Oh Song, Yu Xiang, Stefanie Jegelka, and Silvio Savarese. 2016. Deep metric learning via lifted structured feature embedding. In CVPR. 4004–4012.
[34]
Michael Opitz, Georg Waltner, Horst Possegger, and Horst Bischof. 2017. Bier-boosting independent embeddings robustly. In ICCV. 5189–5198.
[35]
Brandon Shane Parker and Latifur Khan. 2015. Detecting and Tracking Concept Class Drift and Emergence in Non-Stationary Fast Data Streams. In AAAI. 2908–2913.
[36]
Stanislav Pidhorskyi, Ranya Almohsen, and Gianfranco Doretto. 2018. Generative probabilistic novelty detection with adversarial autoencoders. In NeurIPS. 6822–6833.
[37]
Mahardhika Pratama, Choiru Za’in, Andri Ashfahani, Yew Soon Ong, and Weiping Ding. 2019. Automatic construction of multi-layer perceptron network from streaming examples. In CIKM. ACM, 1171–1180.
[38]
Hang Qi, Matthew Brown, and David G Lowe. 2018. Low-shot learning with imprinted weights. In CVPR. 5822–5830.
[39]
Sylvestre-Alvise Rebuffi, Alexander Kolesnikov, Georg Sperl, and Christoph H Lampert. 2017. icarl: Incremental classifier and representation learning. In CVPR. 2001–2010.
[40]
Mengye Ren, Renjie Liao, Ethan Fetaya, and Richard Zemel. 2019. Incremental Few-Shot Learning with Attention Attractor Networks. In NeurIPS. 5276–5286.
[41]
Adam Santoro, Sergey Bartunov, Matthew Botvinick, Daan Wierstra, and Timothy Lillicrap. 2016. One-shot learning with memory-augmented neural networks. arXiv preprint arXiv:1605.06065(2016).
[42]
Florian Schroff, Dmitry Kalenichenko, and James Philbin. 2015. Facenet: A unified embedding for face recognition and clustering. In CVPR. 815–823.
[43]
Lei Shu, Hu Xu, and Bing Liu. 2017. DOC: Deep Open Classification of Text Documents. In EMNLP. Association for Computational Linguistics, 2911–2916.
[44]
Jake Snell, Kevin Swersky, and Richard Zemel. 2017. Prototypical networks for few-shot learning. In NeurIPS. 4077–4087.
[45]
Yi Sun, Ding Liang, Xiaogang Wang, and Xiaoou Tang. 2015. Deepid3: Face recognition with very deep neural networks. arXiv preprint arXiv:1502.00873(2015).
[46]
Flood Sung, Yongxin Yang, Li Zhang, Tao Xiang, Philip HS Torr, and Timothy M Hospedales. 2018. Learning to compare: Relation network for few-shot learning. In CVPR. 1199–1208.
[47]
Oriol Vinyals, Charles Blundell, Timothy Lillicrap, Daan Wierstra, 2016. Matching networks for one shot learning. In NeurIPS. 3630–3638.
[48]
Apoorv Vyas, Nataraj Jammalamadaka, Xia Zhu, Dipankar Das, Bharat Kaul, and Theodore L Willke. 2018. Out-of-distribution detection using an ensemble of self supervised leave-out classifiers. In ECCV. 550–564.
[49]
Apoorv Vyas, Nataraj Jammalamadaka, Xia Zhu, Dipankar Das, Bharat Kaul, and Theodore L Willke. 2018. Out-of-distribution detection using an ensemble of self supervised leave-out classifiers. In ECCV. 550–564.
[50]
Catherine Wah, Steve Branson, Peter Welinder, Pietro Perona, and Serge Belongie. 2011. The caltech-ucsd birds-200-2011 dataset. (2011).
[51]
Jian Wang, Feng Zhou, Shilei Wen, Xiao Liu, and Yuanqing Lin. 2017. Deep metric learning with angular loss. In ICCV. 2593–2601.
[52]
Xiaolong Wang, Abhinav Shrivastava, and Abhinav Gupta. 2017. A-fast-rcnn: Hard positive generation via adversary for object detection. In CVPR. 2606–2615.
[53]
Z. Wang, Z. Kong, S. Changra, H. Tao, and L. Khan. 2019. Robust High Dimensional Stream Classification with Novel Class Detection. In ICDE. 1418–1429.
[54]
Kilian Q Weinberger and Lawrence K Saul. 2009. Distance metric learning for large margin nearest neighbor classification. Journal of Machine Learning Research 10, Feb (2009), 207–244.
[55]
Wei Xia, Jing Huang, and John HL Hansen. 2019. Cross-lingual Text-independent Speaker Verification Using Unsupervised Adversarial Discriminative Domain Adaptation. In ICASSP. IEEE, 5816–5820.
[56]
Han Xiao, Kashif Rasul, and Roland Vollgraf. 2017. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747(2017).
[57]
Eric P Xing, Michael I Jordan, Stuart J Russell, and Andrew Y Ng. 2003. Distance metric learning with application to clustering with side-information. In NeurIPS. 521–528.
[58]
Hu Xu, Bing Liu, Lei Shu, and P Yu. 2019. Open-world Learning and Application to Product Classification. In The World Wide Web Conference. ACM, 3413–3419.
[59]
Zili Yi, Hao Zhang, Ping Tan, and Minglun Gong. 2017. Dualgan: Unsupervised dual learning for image-to-image translation. In ICCV. 2849–2857.
[60]
He Zhang and Vishal M Patel. 2016. Sparse representation-based open set recognition. IEEE transactions on pattern analysis and machine intelligence 39, 8(2016), 1690–1696.

Cited By

View all
  • (2024)A Systematic Literature Review of Novelty Detection in Data Streams: Challenges and OpportunitiesACM Computing Surveys10.1145/365728656:10(1-37)Online publication date: 12-Apr-2024
  • (2023)Big Data Detection utilizing Cloud Networks with Video Vision TechniquesICST Transactions on Scalable Information Systems10.4108/eetsis.vi.3425Online publication date: 8-Jun-2023
  • (2023)Secure Multimedia Data Systems in the Era of Artificial Intelligence: Significant Progress and Vision for the Future2023 5th IEEE International Conference on Trust, Privacy and Security in Intelligent Systems and Applications (TPS-ISA)10.1109/TPS-ISA58951.2023.00031(180-183)Online publication date: 1-Nov-2023
  • Show More Cited By

Index Terms

  1. Few-Sample and Adversarial Representation Learning for Continual Stream Mining
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Please enable JavaScript to view thecomments powered by Disqus.

          Information & Contributors

          Information

          Published In

          cover image ACM Conferences
          WWW '20: Proceedings of The Web Conference 2020
          April 2020
          3143 pages
          ISBN:9781450370233
          DOI:10.1145/3366423
          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Sponsors

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          Published: 20 April 2020

          Permissions

          Request permissions for this article.

          Check for updates

          Author Tags

          1. Adversarial learning
          2. Deep Neural Network
          3. Few Sample Learning
          4. Novel Class Detection
          5. Stream Mining

          Qualifiers

          • Research-article
          • Research
          • Refereed limited

          Conference

          WWW '20
          Sponsor:
          WWW '20: The Web Conference 2020
          April 20 - 24, 2020
          Taipei, Taiwan

          Acceptance Rates

          Overall Acceptance Rate 1,899 of 8,196 submissions, 23%

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • Downloads (Last 12 months)13
          • Downloads (Last 6 weeks)1
          Reflects downloads up to 12 Dec 2024

          Other Metrics

          Citations

          Cited By

          View all
          • (2024)A Systematic Literature Review of Novelty Detection in Data Streams: Challenges and OpportunitiesACM Computing Surveys10.1145/365728656:10(1-37)Online publication date: 12-Apr-2024
          • (2023)Big Data Detection utilizing Cloud Networks with Video Vision TechniquesICST Transactions on Scalable Information Systems10.4108/eetsis.vi.3425Online publication date: 8-Jun-2023
          • (2023)Secure Multimedia Data Systems in the Era of Artificial Intelligence: Significant Progress and Vision for the Future2023 5th IEEE International Conference on Trust, Privacy and Security in Intelligent Systems and Applications (TPS-ISA)10.1109/TPS-ISA58951.2023.00031(180-183)Online publication date: 1-Nov-2023
          • (2022)Meta-segPattern Recognition10.1016/j.patcog.2022.108586126:COnline publication date: 1-Jun-2022
          • (2021)CLEAR: Contrastive-Prototype Learning with Drift Estimation for Resource Constrained Stream MiningProceedings of the Web Conference 202110.1145/3442381.3449820(1351-1362)Online publication date: 19-Apr-2021
          • (2021)CIFDM: Continual and Interactive Feature Distillation for Multi-Label Stream LearningProceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval10.1145/3404835.3463096(2121-2125)Online publication date: 11-Jul-2021
          • (2021)Attention-Based Spatial Guidance for Image-to-Image Translation2021 IEEE Winter Conference on Applications of Computer Vision (WACV)10.1109/WACV48630.2021.00086(816-825)Online publication date: Jan-2021
          • (2021)DEAAN: Disentangled Embedding and Adversarial Adaptation Network for Robust Speaker Representation LearningICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)10.1109/ICASSP39728.2021.9413403(6169-6173)Online publication date: 6-Jun-2021
          • (2020)A Review of Video Object Detection: Datasets, Metrics and MethodsApplied Sciences10.3390/app1021783410:21(7834)Online publication date: 4-Nov-2020
          • (2020)A Primal-Dual Subgradient Approach for Fair Meta Learning2020 IEEE International Conference on Data Mining (ICDM)10.1109/ICDM50108.2020.00091(821-830)Online publication date: Nov-2020

          View Options

          Login options

          View options

          PDF

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          HTML Format

          View this article in HTML Format.

          HTML Format

          Media

          Figures

          Other

          Tables

          Share

          Share

          Share this Publication link

          Share on social media