[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3534678.3539329acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article
Open access

TARNet: Task-Aware Reconstruction for Time-Series Transformer

Published: 14 August 2022 Publication History

Abstract

Time-series data contains temporal order information that can guide representation learning for predictive end tasks (e.g., classification, regression). Recently, there are some attempts to leverage such order information to first pre-train time-series models by reconstructing time-series values of randomly masked time segments, followed by an end-task fine-tuning on the same dataset, demonstrating improved end-task performance. However, this learning paradigm decouples data reconstruction from the end task. We argue that the representations learnt in this way are not informed by the end task and may, therefore, be sub-optimal for the end-task performance. In fact, the importance of different timestamps can vary significantly in different end tasks. We believe that representations learnt by reconstructing important timestamps would be a better strategy for improving end-task performance. In this work, we propose TARNet, Task-Aware Reconstruction Network, a new model using Transformers to learn task-aware data reconstruction that augments end-task performance. Specifically, we design a data-driven masking strategy that uses self-attention score distribution from end-task training to sample timestamps deemed important by the end task. Then, we mask out data at those timestamps and reconstruct them, thereby making the reconstruction task-aware. This reconstruction task is trained alternately with the end task at every epoch, sharing parameters in a single model, allowing the representation learnt through reconstruction to improve end-task performance. Extensive experiments on tens of classification and regression datasets show that TARNet significantly outperforms state-of-the-art baseline models across all evaluation metrics.

Supplemental Material

MP4 File
Presentation video - Reconstructing time-series values of randomly masked time segments learns a self-supervised representation from unlabeled time-series data. This followed by an end-task finetuning has resulted in improved end-task performance. However, this learning paradigm decouples data reconstruction from the end task. These learned representations do not exploit knowledge of the end task, leading to sub-optimal end-task performance. In fact, the importance of different timestamps can vary significantly in different end tasks. We believe that representations learned by reconstructing important timestamps would better improve end-task performance. In this work, we propose TARNet, Task-Aware Reconstruction Network, that trains data reconstruction and the end-task alternately in a multitask setup. We use self-attention score distribution to sample timestamps deemed important by the end task. Then, we mask out data at those timestamps and reconstruct them, thereby making the reconstruction end-task aware.

References

[1]
Anthony Bagnall, Hoang Anh Dau, Jason Lines, Michael Flynn, James Large, Aaron Bostrom, Paul Southam, and Eamonn Keogh. 2018. The UEA multivariate time series classification archive, 2018. arXiv preprint arXiv:1811.00075 (2018).
[2]
Yue Bai, Lichen Wang, Zhiqiang Tao, Sheng Li, and Yun Fu. 2021. Correlative Channel-Aware Fusion for Multi-View Time Series Classification. In AAAI.
[3]
Mustafa Gokce Baydogan and George Runger. 2015. Learning a symbolic representation for multivariate time series classification. Data Mining and Knowledge Discovery 29, 2 (2015), 400--422.
[4]
Tianqi Chen and Carlos Guestrin. 2016. Xgboost: A scalable tree boosting system. In KDD. 785--794.
[5]
Angus Dempster, François Petitjean, and Geoffrey I Webb. 2020. ROCKET: exceptionally fast and accurate time series classification using random convolutional kernels. Data Mining and Knowledge Discovery 34, 5 (2020), 1454--1495.
[6]
Angus Dempster, Daniel F Schmidt, and Geoffrey I Webb. 2021. MINIROCKET: A very fast (almost) deterministic transform for time series classification. In KDD.
[7]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
[8]
Abhilash Dorle, Fangyu Li, Wenzhan Song, and Sheng Li. 2020. Learning Discriminative Virtual Sequences for Time Series Classification. In CIKM. 2001--2004.
[9]
Harris Drucker, Chris JC Burges, Linda Kaufman, Alex Smola, Vladimir Vapnik, et al. 1997. Support vector regression machines. NIPS 9 (1997), 155--161.
[10]
Dheeru Dua and Casey Graff. 2017. UCI Machine Learning Repository. http: //archive.ics.uci.edu/ml
[11]
Emadeldeen Eldele, Mohamed Ragab, Zhenghua Chen, Min Wu, Chee Keong Kwoh, Xiaoli Li, and Cuntai Guan. 2021. Time-Series Representation Learning via Temporal and Contextual Contrasting. arXiv preprint arXiv:2106.14112 (2021).
[12]
Hassan Ismail Fawaz, Germain Forestier, Jonathan Weber, Lhassane Idoumghar, and Pierre-Alain Muller. 2019. Deep learning for time series classification: a review. Data mining and knowledge discovery 33, 4 (2019), 917--963.
[13]
Hassan Ismail Fawaz, Benjamin Lucas, Germain Forestier, Charlotte Pelletier, Daniel F Schmidt, Jonathan Weber, Geoffrey I Webb, Lhassane Idoumghar, PierreAlain Muller, and François Petitjean. 2020. Inceptiontime: Finding alexnet for time series classification. Data Mining and Knowledge Discovery 34, 6 (2020).
[14]
Jean-Yves Franceschi, Aymeric Dieuleveut, and Martin Jaggi. 2019. Unsupervised scalable representation learning for multivariate time series. arXiv preprint arXiv:1901.10738 (2019).
[15]
Dezhi Hong, Jorge Ortiz, Kamin Whitehouse, and David Culler. 2013. Towards automatic spatial verification of sensor placement in buildings. In Proceedings of the 5th ACM Workshop on Embedded Systems For Energy-Efficient Buildings. 1--8.
[16]
Lu Hou, James Kwok, and Jacek Zurada. 2016. Efficient learning of timeseries shapelets. In AAAI, Vol. 30.
[17]
Tsung-Yu Hsieh, Suhang Wang, Yiwei Sun, and Vasant Honavar. 2021. Explainable Multivariate Time Series Classification: A Deep Neural Network Which Learns to Attend to Important Variables As Well As Time Intervals. In WSDM.
[18]
Fazle Karim, Somshubra Majumdar, Houshang Darabi, and Samuel Harford. 2019. Multivariate LSTM-FCNs for time series classification. Neural Networks 116 (2019), 237--245.
[19]
Dongha Lee, Seonghyeon Lee, and Hwanjo Yu. 2021. Learnable Dynamic Temporal Pooling for Time Series Classification. arXiv preprint arXiv:2104.02577 (2021).
[20]
Guozhong Li, Byron Choi, Jianliang Xu, Sourav S Bhowmick, Kwok-Pan Chun, and Grace LH Wong. 2021. Shapenet: A shapelet-neural network approach for multivariate time series classification. In AAAI, Vol. 35. 8375--8383.
[21]
Haoran Liang, Lei Song, Jianxing Wang, Lili Guo, Xuzhi Li, and Ji Liang. 2021. Robust unsupervised anomaly detection via multi-time scale DCGANs with forgetting mechanism for industrial multivariate time series. Neurocomputing 423 (2021), 444--462.
[22]
Qianli Ma, Zhenjing Zheng, Jiawei Zheng, Sen Li, Wanqing Zhuang, and Garrison W Cottrell. 2021. Joint-Label Learning by Dual Augmentation for Time Series Classification. In AAAI, Vol. 35. 8847--8855.
[23]
Qianli Ma, Wanqing Zhuang, Sen Li, Desen Huang, and Garrison Cottrell. 2020. Adversarial dynamic shapelet networks. In AAAI, Vol. 34. 5069--5076.
[24]
Patrick Schäfer and Ulf Leser. 2017. Multivariate time series classification with WEASEL+ MUSE. arXiv preprint arXiv:1711.11343 (2017).
[25]
Mohammad Shokoohi-Yekta, Jun Wang, and Eamonn Keogh. 2015. On the nontrivial generalization of dynamic time warping to the multi-dimensional case. In ICDM. SIAM, 289--297.
[26]
Vladimir Svetnik, Andy Liaw, Christopher Tong, J Christopher Culberson, Robert P Sheridan, and Bradley P Feuston. 2003. Random forest: a classification and regression tool for compound classification and QSAR modeling. Journal of chemical information and computer sciences 43, 6 (2003), 1947--1958.
[27]
Chang Wei Tan, Christoph Bergmeir, Francois Petitjean, and Geoffrey I Webb. 2020. Monash university, uea, ucr time series regression archive. arXiv e-prints (2020), arXiv--2006.
[28]
Sana Tonekaboni, Danny Eytan, and Anna Goldenberg. 2021. Unsupervised representation learning for time series with temporal neighborhood coding. arXiv preprint arXiv:2106.00750 (2021).
[29]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information processing systems. 5998--6008.
[30]
Zhiguang Wang, Weizhong Yan, and Tim Oates. 2017. Time series classification from scratch with deep neural networks: A strong baseline. In 2017 International joint conference on neural networks (IJCNN). IEEE, 1578--1585.
[31]
Xiaopeng Xi, Eamonn Keogh, Christian Shelton, Li Wei, and Chotirat Ann Ratanamahatana. 2006. Fast time series classification using numerosity reduction. In ICML. 1033--1040.
[32]
Chao-Han Huck Yang, Yun-Yun Tsai, and Pin-Yu Chen. 2021. Voice2series: Reprogramming acoustic models for time series classification. In ICML.
[33]
Lexiang Ye and Eamonn Keogh. 2009. Time series shapelets: a new primitive for data mining. In KDD. 947--956.
[34]
Ye Yuan, Guangxu Xun, Fenglong Ma, Yaqing Wang, Nan Du, Kebin Jia, Lu Su, and Aidong Zhang. 2018. Muvan: A multi-view attention network for multivariate temporal data. In ICDM. IEEE, 717--726.
[35]
Zhihan Yue, Yujing Wang, Juanyong Duan, Tianmeng Yang, Congrui Huang, Yunhai Tong, and Bixiong Xu. 2021. TS2Vec: Towards Universal Representation of Time Series. arXiv preprint arXiv:2106.10466 (2021).
[36]
Zhihan Yue, Yujing Wang, Juanyong Duan, Tianmeng Yang, Congrui Huang, and Bixiong Xu. 2021. Learning Timestamp-Level Representations for Time Series with Hierarchical Contrastive Loss. arXiv preprint arXiv:2106.10466 (2021).
[37]
George Zerveas, Srideepika Jayaraman, Dhaval Patel, Anuradha Bhamidipaty, and Carsten Eickhoff. [n. d.]. A Transformer-based Framework for Multivariate Time Series Representation Learning. In KDD, pages=2114--2124, year=2021.
[38]
Xuchao Zhang, Yifeng Gao, Jessica Lin, and Chang-Tien Lu. 2020. Tapnet: Multivariate time series classification with attentional prototypical network. In AAAI.
[39]
Yi Zheng, Qi Liu, Enhong Chen, Yong Ge, and J Leon Zhao. 2014. Time series classification using multi-channels deep convolutional neural networks. In International conference on web-age information management. Springer, 298--310

Cited By

View all
  • (2024)DARKER: Efficient Transformer with Data-Driven Attention Mechanism for Time SeriesProceedings of the VLDB Endowment10.14778/3681954.368199617:11(3229-3242)Online publication date: 30-Aug-2024
  • (2024)A Multi-Scale Decomposition MLP-Mixer for Time Series AnalysisProceedings of the VLDB Endowment10.14778/3654621.365463717:7(1723-1736)Online publication date: 30-May-2024
  • (2024)Time series classification of multi-channel nerve cuff recordings using deep learningPLOS ONE10.1371/journal.pone.029927119:3(e0299271)Online publication date: 12-Mar-2024
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
KDD '22: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
August 2022
5033 pages
ISBN:9781450393850
DOI:10.1145/3534678
This work is licensed under a Creative Commons Attribution International 4.0 License.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 14 August 2022

Check for updates

Author Tags

  1. data reconstruction
  2. self-attention
  3. self-supervision
  4. time series

Qualifiers

  • Research-article

Funding Sources

Conference

KDD '22
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)3,691
  • Downloads (Last 6 weeks)371
Reflects downloads up to 11 Dec 2024

Other Metrics

Citations

Cited By

View all
  • (2024)DARKER: Efficient Transformer with Data-Driven Attention Mechanism for Time SeriesProceedings of the VLDB Endowment10.14778/3681954.368199617:11(3229-3242)Online publication date: 30-Aug-2024
  • (2024)A Multi-Scale Decomposition MLP-Mixer for Time Series AnalysisProceedings of the VLDB Endowment10.14778/3654621.365463717:7(1723-1736)Online publication date: 30-May-2024
  • (2024)Time series classification of multi-channel nerve cuff recordings using deep learningPLOS ONE10.1371/journal.pone.029927119:3(e0299271)Online publication date: 12-Mar-2024
  • (2024)Multi-Hop Multi-View Memory Transformer for Session-Based RecommendationACM Transactions on Information Systems10.1145/366376042:6(1-28)Online publication date: 11-Jul-2024
  • (2024)Deep Learning for Time Series Classification and Extrinsic Regression: A Current SurveyACM Computing Surveys10.1145/364944856:9(1-45)Online publication date: 25-Apr-2024
  • (2024)GinAR: An End-To-End Multivariate Time Series Forecasting Model Suitable for Variable MissingProceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining10.1145/3637528.3672055(3989-4000)Online publication date: 25-Aug-2024
  • (2024)Orthogonality Matters: Invariant Time Series Representation for Out-of-distribution ClassificationProceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining10.1145/3637528.3671768(2674-2685)Online publication date: 25-Aug-2024
  • (2024)Self-Supervised Learning of Time Series Representation via Diffusion Process and Imputation-Interpolation-Forecasting MaskProceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining10.1145/3637528.3671673(2560-2571)Online publication date: 25-Aug-2024
  • (2024)Comparative Analysis of Transformers for Modeling Tabular Data: A Casestudy using Industry Scale DatasetProceedings of the 7th Joint International Conference on Data Science & Management of Data (11th ACM IKDD CODS and 29th COMAD)10.1145/3632410.3632456(449-453)Online publication date: 4-Jan-2024
  • (2024)Self-Supervised Learning for Time Series Analysis: Taxonomy, Progress, and ProspectsIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2024.338731746:10(6775-6794)Online publication date: Oct-2024
  • Show More Cited By

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media