[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

POLLA: Enhancing the Local Structure Awareness in Long Sequence Spatial-temporal Modeling

Published: 29 November 2021 Publication History

Abstract

The spatial-temporal modeling on long sequences is of great importance in many real-world applications. Recent studies have shown the potential of applying the self-attention mechanism to improve capturing the complex spatial-temporal dependencies. However, the lack of underlying structure information weakens its general performance on long sequence spatial-temporal problem. To overcome this limitation, we proposed a novel method, named the Proximity-aware Long Sequence Learning framework, and apply it to the spatial-temporal forecasting task. The model substitutes the canonical self-attention by leveraging the proximity-aware attention, which enhances local structure clues in building long-range dependencies with a linear approximation of attention scores. The relief adjacency matrix technique can utilize the historical global graph information for consistent proximity learning. Meanwhile, the reduced decoder allows for fast inference in a non-autoregressive manner. Extensive experiments are conducted on five large-scale datasets, which demonstrate that our method achieves state-of-the-art performance and validates the effectiveness brought by local structure information.

References

[1]
2019. A deep spatial-temporal data-driven approach considering microclimates for power system security assessment. Appl. Energy 237 (2019), 36–48.
[2]
Sami Abu-El-Haija, Bryan Perozzi, Amol Kapoor, Nazanin Alipourfard, Kristina Lerman, Hrayr Harutyunyan, Greg Ver Steeg, and Aram Galstyan. 2019. MixHop: Higher-order graph convolutional architectures via sparsified neighborhood mixing. In ICML’19, Proceedings of Machine Learning Research, Vol. 97. 21–29.
[3]
Iz Beltagy, Matthew E. Peters, and Arman Cohan. 2020. Longformer: The long-document transformer. arXiv:2004.05150. Retrieved from https://arxiv.org/abs/2004.05150l.
[4]
Salah Bouktif, Ali Fiaz, Ali Ouni, and Mohamed Adel Serhani. 2018. Optimal deep learning LSTM model for electric load forecasting using feature selection and genetic algorithm: Comparison with machine learning approaches. Energies 11, 7 (2018), 1–20.
[5]
Rewon Child, Scott Gray, Alec Radford, and Ilya Sutskever. 2019. Generating long sequences with sparse transformers. arXiv:1904.10509. Retrieved from https://arxiv.org/abs/1904.10509.
[6]
Michaël Defferrard, Xavier Bresson, and Pierre Vandergheynst. 2016. Convolutional neural networks on graphs with fast localized spectral filtering. In NIPS’16. 3837–3845.
[7]
Edouard Delasalles, Ali Ziat, Ludovic Denoyer, and Patrick Gallinari. 2019. Spatio-temporal neural networks for space-time data modeling and relation discovery. Knowl. Inf. Syst. 61, 3 (2019), 1241–1267.
[8]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In NAACL-HLT’19, Volume 1 (Long and Short Papers). Association for Computational Linguistics, 4171–4186.
[9]
Harris Drucker, Christopher J. C. Burges, Linda Kaufman, Alexander J. Smola, and Vladimir Vapnik. 1996. Support vector regression machines. In NIPS’96. MIT Press, 155–161.
[10]
Bowen Du, Hao Peng, Senzhang Wang, Md. Zakirul Alam Bhuiyan, Lihong Wang, Qiran Gong, Lin Liu, and Jing Li. 2020. Deep irregular convolutional residual LSTM for urban traffic passenger flows prediction. IEEE Trans. Intell. Transp. Syst. 21, 3 (2020), 972–985. https://doi.org/10.1109/TITS.2019.2900481
[11]
Jie Feng, Yong Li, Chao Zhang, Funing Sun, Fanchao Meng, Ang Guo, and Depeng Jin. 2018. DeepMove: Predicting human mobility with attentional recurrent networks. In WWW’18. ACM, 1459–1468.
[12]
Alex Graves. 2013. Generating sequences with recurrent neural networks. arXiv:1308.0850. Retrieved from https://arxiv.org/abs/1308.0850.
[13]
Shengnan Guo, Youfang Lin, Shijie Li, Zhaoming Chen, and Huaiyu Wan. 2019. Deep spatial-temporal 3D convolutional neural networks for traffic data forecasting. IEEE Trans. Intell. Transp. Syst. 20, 10 (2019), 3913–3926.
[14]
James Douglas Hamilton. 2020. Time Series Analysis. Princeton University Press.
[15]
Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural Comput. 9, 8 (1997), 1735–1780.
[16]
Chao Huang, Junbo Zhang, Yu Zheng, and Nitesh V. Chawla. 2018. DeepCrime: Attentive hierarchical recurrent networks for crime prediction. In CIKM’18. ACM, 1423–1432.
[17]
Peter J. Huber. 1992. Robust estimation of a location parameter. In Breakthroughs in Statistics. Springer, 492–518.
[18]
Yiannis Kamarianakis and Poulicos Prastacos. 2003. Forecasting traffic flow conditions in an urban network: Comparison of multivariate and univariate approaches. Transport. Res. Rec. 1857, 1 (2003), 74–84.
[19]
Yiannis Kamarianakis and Poulicos Prastacos. 2005. Space-time modeling of traffic flow. Comput. Geosci. 31, 2 (2005), 119–133.
[20]
Antonios Karatzoglou, Nikolai Schnell, and Michael Beigl. 2018. A convolutional neural network approach for modeling semantic trajectories and predicting future locations. In ICANN’18Lecture Notes in Computer Science, Vol. 11139. Springer, 61–72.
[21]
Angelos Katharopoulos, Apoorv Vyas, Nikolaos Pappas, and François Fleuret. 2020. Transformers are RNNs: Fast autoregressive transformers with linear attention. arXiv:2006.16236. Retrieved from https://arxiv.org/abs/2006.16236.
[22]
Seongchan Kim, Seungkyun Hong, Minsu Joh, and Sa-Kwang Song. 2017. DeepRain: ConvLSTM network for precipitation prediction using multichannel radar data. arXiv:1711.02316. Retrieved from https://arxiv.org/abs/1711.02316.
[23]
Thomas N. Kipf and Max Welling. 2017. Semi-supervised classification with graph convolutional networks. In ICLR 2017. OpenReview.net.
[24]
Nikita Kitaev, Lukasz Kaiser, and Anselm Levskaya. 2020. Reformer: The efficient transformer. arXiv:2001.04451. Retrieved from https://arxiv.org/abs/2001.04451.
[25]
Ruoyu Li, Sheng Wang, Feiyun Zhu, and Junzhou Huang. 2018. Adaptive graph convolutional neural networks. In AAAI’18. AAAI Press, 3546–3553.
[26]
Shiyang Li, Xiaoyong Jin, Yao Xuan, Xiyou Zhou, Wenhu Chen, Yu-Xiang Wang, and Xifeng Yan. 2019. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. In NIPS’19. 5244–5254.
[27]
Yaguang Li, Rose Yu, Cyrus Shahabi, and Yan Liu. 2018. Diffusion convolutional recurrent neural network: Data-driven traffic forecasting. In ICLR’18. OpenReview.net.
[28]
Qiang Liu, Shu Wu, Liang Wang, and Tieniu Tan. 2016. Predicting the next location: A recurrent model with spatial and temporal contexts. In AAAI’16. AAAI Press, 194–200.
[29]
Zhongjian Lv, Jiajie Xu, Kai Zheng, Hongzhi Yin, Pengpeng Zhao, and Xiaofang Zhou. 2018. LC-RNN: A deep learning model for traffic speed prediction. In IJCAI’18. 3470–3476.
[30]
Li Mengzhang and Zhu Zhanxing. 2020. Spatial-temporal fusion graph neural networks for traffic flow forecasting. arXiv:2012.09641 [cs.LG]. Retrieved from https://arxiv.org/abs/2012.09641.
[31]
Cheonbok Park, Chunggi Lee, Hyojin Bahng, Yunwon Tae, Seungmin Jin, Kihwan Kim, Sungahn Ko, and Jaegul Choo. 2020. ST-GRAT: A novel spatio-temporal graph attention networks for accurately forecasting dynamically changing road speed. In CIKM’20. ACM, 1215–1224.
[32]
Cheonbok Park, Chunggi Lee, Hyojin Bahng, Taeyun won, Kihwan Kim, Seungmin Jin, Sungahn Ko, and Jaegul Choo. 2019. STGRAT: A spatio-temporal graph attention network for traffic forecasting. arXiv:1911.13181. Retrieved from https://arxiv.org/abs/1911.13181.
[33]
Hao Peng, Hongfei Wang, Bowen Du, Md. Zakirul Alam Bhuiyan, Hongyuan Ma, Jianwei Liu, Lihong Wang, Zeyu Yang, Linfeng Du, Senzhang Wang, and Philip S. Yu. 2020. Spatial temporal incidence dynamic graph neural networks for traffic flow forecasting. Inf. Sci. 521 (2020), 277–290.
[34]
Chiara Plizzari, Marco Cannici, and Matteo Matteucci. 2020. Spatial temporal transformer network for skeleton-based action recognition. arXiv:2008.07404. Retrieved from https://arxiv.org/abs/2008.07404.
[35]
Xingjian Shi, Zhourong Chen, Hao Wang, Dit-Yan Yeung, Wai-Kin Wong, and Wang-chun Woo. 2015. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In NIPS’15. 802–810.
[36]
David I. Shuman, Sunil K. Narang, Pascal Frossard, Antonio Ortega, and Pierre Vandergheynst. 2013. The emerging field of signal processing on graphs: Extending high-dimensional data analysis to networks and other irregular domains. IEEE Sign. Process. Mag. 30, 3 (2013), 83–98.
[37]
Karen Simonyan and Andrew Zisserman. 2014. Two-stream convolutional networks for action recognition in videos. In NIPS’14. 568–576.
[38]
Chao Song, Youfang Lin, Shengnan Guo, and Huaiyu Wan. 2020. Spatial-temporal synchronous graph convolutional networks: A new framework for spatial-temporal network data forecasting. In AAAI’20. AAAI Press, 914–921.
[39]
Sean J Taylor and Benjamin Letham. 2018. Forecasting at scale. Am. Stat. 72, 1 (2018), 37–45.
[40]
Yao-Hung Hubert Tsai, Shaojie Bai, Makoto Yamada, Louis-Philippe Morency, and Ruslan Salakhutdinov. 2019. Transformer dissection: An unified understanding for transformer’s attention via the lens of kernel. In EMNLP-IJCNLP’19. Association for Computational Linguistics, 4344–4353.
[41]
Aäron van den Oord, Sander Dieleman, Heiga Zen, Karen Simonyan, Oriol Vinyals, Alex Graves, Nal Kalchbrenner, Andrew W. Senior, and Koray Kavukcuoglu. 2016. WaveNet: A generative model for raw audio. In Proceedings of the 9th ISCA Speech Synthesis Workshop. ISCA, 125.
[42]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In NIPS’17. 5998–6008.
[43]
Bao Wang, Xiyang Luo, Fangbo Zhang, Baichuan Yuan, Andrea L. Bertozzi, and P. Jeffrey Brantingham. 2018. Graph-based deep modeling and real time forecasting of sparse spatio-temporal data. arXiv:1804.00684. Retrieved from https://arxiv.org/abs/1804.00684.
[44]
Leye Wang, Xu Geng, Xiaojuan Ma, Feng Liu, and Qiang Yang. 2018. Crowd flow prediction by deep spatio-temporal transfer learning. arXiv:1802.00386. Retrieved from https://arxiv.org/abs/1802.00386.
[45]
Senzhang Wang, Jiannong Cao, and Philip S. Yu. 2019. Deep learning for spatio-temporal data mining: A survey. arXiv:1906.04928. Retrieved from https://arxiv.org/abs/1906.04928.
[46]
Sinong Wang, Belinda Z. Li, Madian Khabsa, Han Fang, and Hao Ma. 2020. Linformer: Self-attention with linear complexity. arXiv:2006.04768. Retrieved from https://arxiv.org/abs/2006.04768.
[47]
Billy M Williams and Lester A Hoel. 2003. Modeling and forecasting vehicular traffic flow as a seasonal ARIMA process: Theoretical basis and empirical results. J. Transport. Eng. 129, 6 (2003), 664–672.
[48]
Zonghan Wu, Shirui Pan, Guodong Long, Jing Jiang, and Chengqi Zhang. 2019. Graph WaveNet for deep spatial-temporal graph modeling. In IJCAI’19. 1907–1913.
[49]
Sijie Yan, Yuanjun Xiong, and Dahua Lin. 2018. Spatial temporal graph convolutional networks for skeleton-based action recognition. In AAAI’18. AAAI Press, 7444–7452.
[50]
Zhixian Yan. 2010. Traj-ARIMA: A spatial-time series model for network-constrained trajectory. In IWCTS’21. ACM, 11–16.
[51]
Bing Yu, Haoteng Yin, and Zhanxing Zhu. 2018. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. In IJCAI’18. 3634–3640.
[52]
Zhuoning Yuan, Xun Zhou, and Tianbao Yang. 2018. Hetero-ConvLSTM: A deep learning approach to traffic accident prediction on heterogeneous spatio-temporal data. In KDD’18. ACM, 984–992.
[53]
Y. Zhang, S. Wang, B. Chen, J. Cao, and Z. Huang. 2021. TrafficGAN: Network-scale deep traffic prediction with generative adversarial nets. IEEE Trans. Intell. Transport. Syst. 22, 1 (2021), 219–230.
[54]
Chuanpan Zheng, Xiaoliang Fan, Cheng Wang, and Jianzhong Qi. 2020. GMAN: A graph multi-attention network for traffic prediction. In AAAI’20. AAAI Press, 1234–1241.
[55]
Chuanpan Zheng, Xiaoliang Fan, Chenglu Wen, Longbiao Chen, Cheng Wang, and Jonathan Li. 2020. DeepSTD: Mining spatio-temporal disturbances of multiple context factors for citywide traffic flow prediction. IEEE Trans. Intell. Transport. Syst. 21, 9 (2020), 3744–3755.
[56]
Haoyi Zhou, Shanghang Zhang, Jieqi Peng, Shuai Zhang, Jianxin Li, Hui Xiong, and Wancai Zhang. 2021. Informer: Beyond efficient transformer for long sequence time-series forecasting. In AAAI’21. AAAI Press.

Cited By

View all
  • (2024)Time-aware personalized graph convolutional network for multivariate time series forecastingExpert Systems with Applications10.1016/j.eswa.2023.122471240(122471)Online publication date: Apr-2024
  • (2024)Dynamic personalized graph neural network with linear complexity for multivariate time series forecastingEngineering Applications of Artificial Intelligence10.1016/j.engappai.2023.107291127:PAOnline publication date: 1-Feb-2024
  • (2022)AIQoSer: Building the efficient Inference-QoS for AI Services2022 IEEE/ACM 30th International Symposium on Quality of Service (IWQoS)10.1109/IWQoS54832.2022.9812905(1-10)Online publication date: 10-Jun-2022
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Transactions on Intelligent Systems and Technology
ACM Transactions on Intelligent Systems and Technology  Volume 12, Issue 6
December 2021
356 pages
ISSN:2157-6904
EISSN:2157-6912
DOI:10.1145/3501281
  • Editor:
  • Huan Liu
Issue’s Table of Contents

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 29 November 2021
Accepted: 01 January 2021
Revised: 01 December 2020
Received: 01 October 2020
Published in TIST Volume 12, Issue 6

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Spatio-temporal
  2. neural network
  3. graph
  4. time-series

Qualifiers

  • Research-article
  • Refereed

Funding Sources

  • Natural Science Foundation of China
  • State Key Laboratory of Software Development Environment
  • CAAI-Huawei MindSpore Open Fund

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)101
  • Downloads (Last 6 weeks)8
Reflects downloads up to 13 Dec 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Time-aware personalized graph convolutional network for multivariate time series forecastingExpert Systems with Applications10.1016/j.eswa.2023.122471240(122471)Online publication date: Apr-2024
  • (2024)Dynamic personalized graph neural network with linear complexity for multivariate time series forecastingEngineering Applications of Artificial Intelligence10.1016/j.engappai.2023.107291127:PAOnline publication date: 1-Feb-2024
  • (2022)AIQoSer: Building the efficient Inference-QoS for AI Services2022 IEEE/ACM 30th International Symposium on Quality of Service (IWQoS)10.1109/IWQoS54832.2022.9812905(1-10)Online publication date: 10-Jun-2022
  • (2021)An Effective Segmentation of Tissues from MR Brain ImagesJournal of Physics: Conference Series10.1088/1742-6596/1964/6/0620291964:6(062029)Online publication date: 1-Jul-2021

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Full Text

View this article in Full Text.

Full Text

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media