Unveiling Taxi Drivers' Strategies via cGAIL: Conditional Generative Adversarial Imitation Learning

Smart passenger-seeking strategies employed by taxi drivers contribute not only to drivers' incomes, but also higher quality of service passengers received. Therefore, understanding taxi drivers' behaviors and learning the good passenger-seeking strategies are crucial to boost taxi drivers' well-being and public transportation quality of service. However, we observe that drivers' preferences of choosing which area to find the next passenger are diverse and dynamic across locations and drivers. It is hard to learn the location-dependent preferences given the partial data (i.e., an individual driver's trajectory may not cover all locations). In this paper, we make the first attempt to develop conditional generative adversarial imitation learning (cGAIL) model, as a unifying collective inverse reinforcement learning framework that learns the driver's decision-making preferences and policies by transferring knowledge across taxi driver agents and across locations. Our evaluation results on three months of taxi GPS trajectory data in Shenzhen, China, demonstrate that the driver's preferences and policies learned from cGAIL are on average 34.7% more accurate than those learned from other state-of-the-art baseline approaches.

[1]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[2]  Yu Zheng,et al.  Human-Centric Urban Transit Evaluation and Planning , 2018, 2018 IEEE International Conference on Data Mining (ICDM).

[3]  Jan Peters,et al.  Relative Entropy Inverse Reinforcement Learning , 2011, AISTATS.

[4]  Brett Browning,et al.  A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..

[5]  Stefano Ermon,et al.  Generative Adversarial Imitation Learning , 2016, NIPS.

[6]  Anind K. Dey,et al.  Maximum Entropy Inverse Reinforcement Learning , 2008, AAAI.

[7]  Fan Zhang,et al.  Growing the charging station network for electric vehicles with trajectory data analytics , 2015, 2015 IEEE 31st International Conference on Data Engineering.

[8]  Sergey Levine,et al.  Learning Robust Rewards with Adversarial Inverse Reinforcement Learning , 2017, ICLR 2017.

[9]  Huaiyu Zhu On Information and Sufficiency , 1997 .

[10]  Andrew Y. Ng,et al.  Pharmacokinetics of a novel formulation of ivermectin after administration to goats , 2000, ICML.

[11]  Pieter Abbeel,et al.  Apprenticeship learning via inverse reinforcement learning , 2004, ICML.

[12]  Lawrence D. Jackel,et al.  Handwritten Digit Recognition with a Back-Propagation Network , 1989, NIPS.

[13]  H. Anton,et al.  Elementary linear algebra : applications version , 2008 .

[14]  Fan Zhang,et al.  Data-driven inverse learning of passenger preferences in urban public transits , 2017, 2017 IEEE 56th Annual Conference on Decision and Control (CDC).

[15]  Changhyun Kwon,et al.  Inferring origin-destination pairs and utility-based travel preferences of shared mobility system users in a multi-modal environment , 2016 .

[16]  Anind K. Dey,et al.  Modeling Interaction via the Principle of Maximum Causal Entropy , 2010, ICML.

[17]  Ting Zhu,et al.  Region sampling and estimation of geosocial data with dynamic range calibration , 2014, 2014 IEEE 30th International Conference on Data Engineering.