Multi-Task Spatial-Temporal Graph Attention Network for Taxi Demand Prediction

Taxi demand prediction is of much importance, which enables the building of intelligent systems and smart city. It is necessary to predict taxi demand accurately to schedule taxi fleet in a reasonable and efficient way and to reduce the pressure of traffic jam. However, the taxi demand involves complex and non-linear spatial-temporal impacts. The superiority of deep learning makes people explore the possibility to apply it to traffic prediction. State-of-the-art methods on taxi demand prediction only capture static spatial correlations between regions (e.g., Using static graph embedding) and only take taxi demand data into consideration. We propose a Multi-Task Spatial-Temporal Graph Attention Network (MSTGAT-Net) framework which models the correlations between regions dynamically with graph-attention network and captures the correlation between taxi pick up and taxi drop off with multi-task training. To the best of our knowledge, it is the first paper to address the taxi demand prediction problem with graph attention network and multi-task learning. Experiments on real-world taxi data show that our model is superior to state-of-the-art methods.

[1]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[2]  Yu Zheng,et al.  Deep Spatio-Temporal Residual Networks for Citywide Crowd Flows Prediction , 2016, AAAI.

[3]  Yann LeCun,et al.  Deep learning & convolutional networks , 2015, 2015 IEEE Hot Chips 27 Symposium (HCS).

[4]  João Gama,et al.  Predicting Taxi–Passenger Demand Using Streaming Data , 2013, IEEE Transactions on Intelligent Transportation Systems.

[5]  Jieping Ye,et al.  Deep Multi-View Spatial-Temporal Network for Taxi Demand Prediction , 2018, AAAI.

[6]  Wei Cao,et al.  DeepSD: Supply-Demand Prediction for Online Car-Hailing Services Using Deep Neural Networks , 2017, 2017 IEEE 33rd International Conference on Data Engineering (ICDE).

[7]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[8]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[9]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[10]  Mingzhe Wang,et al.  LINE: Large-scale Information Network Embedding , 2015, WWW.

[11]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[12]  G. Box,et al.  Distribution of Residual Autocorrelations in Autoregressive-Integrated Moving Average Time Series Models , 1970 .

[13]  Zheng Wang,et al.  Multi-task Representation Learning for Travel Time Estimation , 2018, KDD.

[14]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.