References(41)
[1]
Z. H. Duan, H. Y. Xu, Y. Y. Wang, Y. D. Huang, A. N. Ren, Z. B. Xu, Y. Z. Sun, and W. Wang, Multivariate time series classification with hierarchical variational graph pooling, arXiv preprint arXiv: 2010.05649, 2020.
[2]
H. Chen, S. Feng, X. Pei, Z. Zhang, and D. Yao, Dangerous driving behavior recognition and prevention using an autoregressive time-series model, Tsinghua Science and Technology, vol. 22, no. 6, pp. 682–690, 2017.
[3]
Y. S. Liu, C. H. Yang, K. K. Huang, and W. H. Gui, Nonferrous metals price forecasting based on variational mode decomposition and LSTM network, Knowl. Based Syst., vol. 188, p. 105006, 2020.
[4]
L. Yu, S. Dong, M. Lu, and J. Wang, LSTM based reserve prediction for bank outlets, Tsinghua Science and Technology, 2018, vol. 24, no. 1, pp. 77–85, 2018.
[5]
Y. Y. Wang, Z. H. Duan, Y. D. Huang, H. Y. Xu, J. Feng, and A. N. Ren, MTHetGNN: A heterogeneous graph embedding framework for multivariate time series forecasting, Pattern Recognit. Lett., vol. 153, pp. 151–158, 2022.
[6]
G. E. P. Box, G. M. Jenkins, G. C. Reinsel, and G. M. Ljung, Time Series Analysis: Forecasting and Control, 4th Edition, Hoboken, NJ, USA: John Wiley & Sons, 2016.
[7]
J. D. Hamilton, Time Series Analysis. Princeton, NJ, USA: Princeton University Press, 1994.
[8]
H. Lütkepohl, New Introduction to Multiple Time Series Analysis. Berlin, Germany: Springer, 2005.
[9]
A. Tokgöz and G. Ünal, A RNN based time series approach for forecasting Turkish electricity load, in Proc. 2018 26th Signal Processing and Communications Applications Conf. (SIU), Izmir, Turkey, 2018, pp. 1–4.
[10]
J. L. Elman, Finding structure in time, Cogn. Sci., vol. 14, no. 2, pp. 179–211, 1990.
[11]
S. Hochreiter and J. Schmidhuber, Long short-term memory, Neural Comput., vol. 9, no. 8, pp. 1735–1780, 1997.
[12]
J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, Empirical evaluation of gated recurrent neural networks on sequence modeling, arXiv preprint arXiv: 1412.3555, 2014.
[13]
G. K. Lai, W. C. Chang, Y. M. Yang, and H. X. Liu, Modeling long- and short-term temporal patterns with deep neural networks, in Proc. the 41st Int. ACM SIGIR Conf. on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 2017, pp. 95–104.
[14]
Y. F. Zhou, Z. H. Duan, H. Y. Xu, J. Feng, A. N. Ren, Y. Y. Wang, and X. Q. Wang, Parallel extraction of long-term trends and short-term fluctuation framework for multivariate time series forecasting, arXiv preprint arXiv: 2008.07730, 2020.
[15]
C. W. J. Granger, Investigating causal relations by econometric models and cross-spectral methods, Econometrica, vol. 37, no. 3, pp. 424–438, 1969.
[16]
G. Kirchgässner, J. Wolters, and U. Hassler, Granger causality, in Introduction to Modern Time Series Analysis, G. Kirchgässner, J. Wolters, and U. Hassler, eds. Berlin, Germany: Springer, 2013, pp. 95–125.
[17]
T. Bossomaier, L. Barnett, M. Harré, and J. T. Lizier, Transfer entropy, in An Introduction to Transfer Entropy, T. Bossomaier, L. Barnett, M. Harré, and J. T. Lizier, eds. Cham, Switzerland: Springer, 2016, pp. 65–95.
[18]
T. Dimpfl and F. J. Peter, Using transfer entropy to measure information flows between financial markets, Stud. Nonlinear Dyn. Econom., vol. 17, no. 1, pp. 85–102, 2013.
[19]
T. Q. Tung, T. Ryu, K. H. Lee, and D. Lee, Inferring gene regulatory networks from microarray time series data using transfer entropy, in Proc. 20th IEEE Int. Symp. on Computer-Based Medical Systems (CBMS’07), Maribor, Slovenia, 2007, pp. 383–388.
[20]
M. Bauer, J. W. Cox, M. H. Caveness, J. J. Downs, and N. F. Thornhill, Finding the direction of disturbance propagation in a chemical process using transfer entropy, IEEE Trans. Control Syst. Technol., vol. 15, no. 1, pp. 12–21, 2007.
[21]
F. Scarselli, M. Gori, A. C. Tsoi, M. Hagenbuchner, and G. Monfardini, The graph neural network model, IEEE Trans. Neural Netw., vol. 20, no. 1, pp. 61–80, 2009.
[22]
Y. Y Wang, Z. H. Duan, B. B. Liao, F. Wu, and Y. T. Zhuang, Heterogeneous attributed network embedding with graph convolutional networks, in Proc. the AAAI Conf. on Artificial Intelligence, Palo Alto, CA, USA, 2019, pp. 10061&10062.
[23]
Z. H. Duan, Y. Y. Wang, W. H. Ye, Z. X. Feng, Q. L. Fan, and X. H. Li, Connecting latent relationships over heterogeneous attributed network for recommendation, arXiv preprint arXiv: 2103.05749, 2021.
[24]
J. Gilmer, S. S. Schoenholz, P. F. Riley, O. Vinyals, and G. E. Dahl, Neural message passing for quantum chemistry, in Proc. the 34th Int. Conf. on Machine Learning, Sydney, Australia, 2017, pp. 1263–1272.
[25]
Z. T. Ying, J. X. You, C. Morris, X. Ren, W. Hamilton, and J. Leskovec, Hierarchical graph representation learning with differentiable pooling, in Proc. Advances in Neural Information Processing Systems, Montréal, Canada, 2018, pp. 4800–4810.
[26]
H. Y. Xu, R. J. Chen, Y. S. Bai, Z. H. Duan, J. Feng, Y. Z. Sun, and W. Wang, CoSimGNN: Towards large-scale graph similarity computation, arXiv preprint arXiv: 2005.07115, 2020.
[27]
H. Y. Xu, Z. H. Duan, Y. Y. Wang, J. Feng, R. J. Chen, Q. R. Zhang, and Z. B. Xu, Graph partitioning and graph neural network based hierarchical graph matching for graph similarity computation, Neurocomputing, vol. 439, pp. 348–362, 2021.
[28]
T. N. Kipf and M. Welling, Semi-supervised classification with graph convolutional networks, arXiv preprint arXiv: 1609.02907, 2016.
[29]
W. L. Hamilton, R. Ying, and J. Leskovec, Inductive representation learning on large graphs, in Proc. the 31st Int. Conf. on Neural Information Processing Systems, Long Beach, CA, USA, 2017, pp. 1025–1035.
[30]
J. Chen, T. F. Ma, and C. Xiao, FastGCN: Fast learning with graph convolutional networks via importance sampling, arXiv preprint arXiv: 1801.10247, 2018.
[31]
K. Y. L. Xu, W. H. Hu, J. Leskovec, and S. Jegelka, How powerful are graph neural networks? arXiv preprint arXiv: 1810.00826, 2018.
[32]
C. Morris, M. Ritzert, M. Fey, W. L. Hamilton, J. E. Lenssen, G. Rattan, and M. Grohe, Weisfeiler and leman go neural: Higher-order graph neural networks, in Proc. the AAAI Conf. on Artificial Intelligence, Palo Alto, CA, USA, 2019, pp. 4602–4609.
[33]
P. Velikovi, G. Cucurull, A. Casanova, A. Romero, P. Liò, and Y. Bengio, Graph attention networks, arXiv preprint arXiv: 1710.10903, 2017.
[34]
S. Han, H. B. Dong, X. Y. Teng, X. H. Li, and X. W. Wang, Correlational graph attention-based long short-term memory network for multivariate time series prediction, Appl. Soft Comput., vol. 106, p. 107377, 2021.
[35]
T. Schreiber, Measuring information transfer, Phys. Rev. Lett., vol. 85, no. 2, pp. 461–464, 2000.
[36]
D. P. Kingma and J. Ba, Adam: A method for stochastic optimization, arXiv preprint arXiv: 1412.6980, 2014.
[37]
L. M. Candanedo, V. Feldheim, and D. Deramaix, Data driven prediction models of energy use of appliances in a low-energy house, Energy Build., vol. 140, pp. 81–97, 2017.
[38]
Y. Qin, D. J. Song, H. F. Cheng, W. Cheng, G. F. Jiang, and G. W. Cottrell, A dual-stage attention-based recurrent neural network for time series prediction, in Proc. the 26th Int. Joint Conf. on Artificial Intelligence, Melbourne, Australia, 2017, pp. 2627–2633.
[39]
Y. LeCun and Y. Bengio, Convolutional networks for images, speech, and time series, in The Handbook of Brain Theory and Neural Networks, M. A. Arbib, ed. Cambridge, MA, USA: MIT Press, 1998, pp. 255–258.
[40]
A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, Attention is all you need, in Proc. the 31st Int. Conf. on Neural Information Processing Systems, Long Beach, CA, USA, 2017, pp. 6000–6010.
[41]
J. Z. Cheng, K. Z. Huang, and Z. B. Zheng, Towards better forecasting by fusing near and distant future visions, in Proc. the AAAI Conf. on Artificial Intelligence, Palo Alto, CA, USA, 2020, pp. 3593–3600.