Skip to main content
Log in

GTHP: a novel graph transformer Hawkes process for spatiotemporal event prediction

  • Regular Paper
  • Published:
Knowledge and Information Systems Aims and scope Submit manuscript

Abstract

The event sequences with spatiotemporal characteristics have been rapidly produced in various domains, such as earthquakes in seismology, electronic medical records in healthcare, and transactions in the financial market. These data often continue for weeks, months, or years, and the past events may trigger subsequent events. In this context, modeling the spatiotemporal event sequences and forecasting the next event has become a hot topic. However, existing models either failed to capture the long-term temporal dependencies or ignored the essential spatial information between sequences. In this paper, we proposed a novel graph transformer Hawkes process (GTHP) model to capture the long-term temporal dependencies and spatial information from historical events. The core concept of GTHP is to learn the spatial information by graph convolutional neural networks and capture long-term temporal dependencies from events embedding by self-attention mechanism. Moreover, we integrated the learned spatial information into the event embedding as auxiliary information. Numerous experiments on synthetic and real-world datasets proved the effectiveness of the proposed model.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Algorithm 1
Algorithm 2

Similar content being viewed by others

Data Availability

The data supporting this study’s findings are available from the corresponding author upon reasonable request.

References

  1. Hawkes AG, Oakes D (1974) A cluster process representation of a self-exciting process. J Appl Probab 11(3):493–503

    Article  MathSciNet  Google Scholar 

  2. Du N, Wang Y, He N, Sun J, Song L (2015) Time-sensitive recommendation from recurrent user activities. Advances in neural information processing systems, vol 28

  3. Yan J, Wang Y, Zhou K, Huang J, Tian C, Zha H, Dong W (2013) Towards effective prioritizing water pipe replacement and rehabilitation. In: Twenty-third international joint conference on artificial intelligence. Citeseer

  4. Ertekin Ş, Rudin C, McCormick TH (2015) Reactive point processes: a new approach to predicting power failures in underground electrical systems. Ann Appl Stat 9(1):122–144

    Article  MathSciNet  Google Scholar 

  5. Du N, Dai H, Trivedi R, Upadhyay U, Gomez-Rodriguez M, Song L (2016) Recurrent marked temporal point processes: Embedding event history to vector. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pp 1555–1564

  6. Xiao S, Yan J, Yang X, Zha H, Chu SM (2017) Modeling the intensity function of point process via recurrent neural networks. In: Thirty-First AAAI conference on artificial intelligence

  7. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780

    Article  CAS  PubMed  Google Scholar 

  8. Chung J, Gulcehre C, Cho K, Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555

  9. Pascanu R, Mikolov T, Bengio Y (2013) On the difficulty of training recurrent neural networks. In: International conference on machine learning. PMLR, pp 1310–1318

  10. Embrechts P, Liniger T, Lin L (2011) Multivariate Hawkes processes: an application to financial data. J Appl Probab 48(A):367–378

    Article  MathSciNet  Google Scholar 

  11. Hawkes AG (1971) Spectra of some self-exciting and mutually exciting point processes. Biometrika 58(1):83–90

    Article  MathSciNet  Google Scholar 

  12. Ogata Y (1988) Statistical models for earthquake occurrences and residual analysis for point processes. J Am Stat Assoc 83(401):9–27

    Article  Google Scholar 

  13. Li L, Zha H (2013) Dyadic event attribution in social networks with mixtures of hawkes processes. In: Proceedings of the 22nd ACM international conference on information & knowledge management, pp 1667–1672

  14. Kobayashi R, Lambiotte R (2016) Tideh: Time-dependent Hawkes process for predicting retweet dynamics. In: Tenth international AAAI conference on web and social media

  15. Li L, Deng H, Dong A, Chang Y, Zha H (2014) Identifying and labeling search tasks via query-based Hawkes processes. In: Proceedings of the 20th ACM SIGKDD international conference on knowledge discovery and data mining, pp 731–740

  16. Li L, Zha H (2016) Household structure analysis via Hawkes processes for enhancing energy disaggregation. In: IJCAI, pp 2553–2559

  17. Salehi F, Trouleau W, Grossglauser M, Thiran P (2019) Learning Hawkes processes from a handful of events. In: Advances in neural information processing systems, pp 12694–12704

  18. Chiang W-H, Liu X, Mohler G (2022) Hawkes process modeling of Covid-19 with mobility leading indicators and spatial covariates. Int J Forecast 38(2):505–520

    Article  PubMed  Google Scholar 

  19. Mei H, Eisner JM (2017) The neural Hawkes process: a neurally self-modulating multivariate point process. In: Advances in neural information processing systems, pp 6754–6764

  20. Liu Y, Yan T, Chen H (2018) Exploiting graph regularized multi-dimensional Hawkes processes for modeling events with spatio-temporal characteristics. In: IJCAI, pp 2475–2482

  21. Shang J, Sun M (2019) Geometric Hawkes processes with graph convolutional recurrent neural networks. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 4878–4885

  22. Zhang Y, Yan J (2021) Neural relation inference for multi-dimensional temporal point processes via message passing graph. In: IJCAI, pp 3406–3412

  23. Du H, Zhou Y, Ma Y, Wang S (2021) Astrologer: exploiting graph neural Hawkes process for event propagation prediction with spatio-temporal characteristics. Knowl-Based Syst 228:107247

    Article  Google Scholar 

  24. Du H, Zhou Y (2022) Nostradamus: a novel event propagation prediction approach with spatio-temporal characteristics in non-Euclidean space. Neural Netw 145:386–394

    Article  PubMed  Google Scholar 

  25. Türkmen AC, Wang Y, Smola AJ (2019) Fastpoint: scalable deep point processes. In: Joint European conference on machine learning and knowledge discovery in databases

  26. Zhang Q, Lipani A, Kirnap O, Yilmaz E (2020) Self-attentive Hawkes process. In: International conference on machine learning. PMLR, pp 11183–11193

  27. Zuo S, Jiang H, Li Z, Zhao T, Zha H (2020) Transformer Hawkes process. In: International conference on machine learning. PMLR, pp 11692–11702

  28. Kipf TN, Welling M (2016) Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907

  29. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. Advances in neural information processing systems, vol 30

  30. Xu H, Farajtabar M, Zha H (2016) Learning granger causality for Hawkes processes. In: International conference on machine learning. PMLR, pp 1717–1726

  31. Rasmussen JG (2011) Temporal point processes: the conditional intensity function. Lecture Notes

  32. Li Y, Du N, Bengio S (2017) Time-dependent representation for neural event sequence prediction. arXiv preprint arXiv:1708.00065

  33. Kingma DP, Ba J (2014) Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980

  34. Ba JL, Kiros JR, Hinton GE (2016) Layer normalization. arXiv preprint arXiv:1607.06450

  35. Yan J, Liu X, Shi L, Li C, Zha H (2018) Improving maximum likelihood estimation of temporal point process via discriminative and adversarial learning. In: IJCAI, pp 2948–2954

  36. Zhou K, Zha H, Song L (2013) Learning triggering kernels for multi-dimensional Hawkes processes. In: International conference on machine learning. PMLR, pp 1301–1309

  37. Grant S, Betts B (2013) Encouraging user behaviour with achievements: an empirical study. In: 2013 10th working conference on mining software repositories (MSR). IEEE, pp 65–68

  38. Monti F, Bronstein M, Bresson X (2017) Geometric matrix completion with recurrent multi-graph neural networks. Advances in neural information processing systems, vol 30

Download references

Author information

Authors and Affiliations

Authors

Contributions

Yiman Xie contributed to the conception of the study; Jianbin Wu contributed to the conception of the study; Yiman Xie performed the experiment; Yiman Xie contributed significantly to analysis and manuscript preparation; Jianbin Wu contributed significantly to analysis and manuscript preparation; Yiman Xie performed the data analyses and wrote the manuscript; Jianbin Wu helped perform the analysis with constructive discussions. Yan Zhou contributed substantially to the paper’s ablation study, experimentation, and revision.

Corresponding author

Correspondence to Jianbin Wu.

Ethics declarations

Conflict of interest

We declare that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work. There is no professional or other personal interest of any nature or kind in any product, service and/or company that could be construed as influencing the position presented in, or the review of, the manuscript entitled.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Xie, Y., Wu, J. & Zhou, Y. GTHP: a novel graph transformer Hawkes process for spatiotemporal event prediction. Knowl Inf Syst (2024). https://doi.org/10.1007/s10115-024-02080-z

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10115-024-02080-z

Keywords

Navigation