Modeling learning and adaptation processes in activity-travel choice A framework and numerical experiment

This paper develops a framework for modeling dynamic choice based on a theory of reinforcement learning and adaptation. According to this theory, individuals develop and continuously adapt choice rules while interacting with their environment. The proposed model framework specifies required components of learning systems including a reward function, incremental action value functions, and action selection methods. Furthermore, the system incorporates an incremental induction method that identifies relevant states based on reward distributions received in the past. The system assumes multi-stage decision making in potentially very large condition spaces and can deal with stochastic, non-stationary, and discontinuous reward functions. A hypothetical case is considered that combines route, destination, and mode choice for an activity under time-varying conditions of the activity schedule and road congestion probabilities. As it turns out, the system is quite robust for parameter settings and has good face validity. We therefore argue that it provides a useful and comprehensive framework for modeling learning and adaptation in the area of activity-travel choice.

[1]  Hani S. Mahmassani,et al.  Experiments with departure time choice dynamics of urban commuters , 1986 .

[2]  Moshe Ben-Akiva,et al.  Dynamic network models and driver information systems , 1991 .

[3]  Satoshi Fujii,et al.  Drivers’ Learning and Network Behavior: Dynamic Analysis of the Driver-Network System as a Complex System , 1999 .

[4]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 2005, IEEE Transactions on Neural Networks.

[5]  Yasunori Iida,et al.  Experimental analysis of dynamic route choice behavior , 1992 .

[6]  Leo Breiman,et al.  Classification and Regression Trees , 1984 .

[7]  Ryuichi Kitamura,et al.  Route Choice Model with Inductive Learning , 2000 .

[8]  G. V. Kass An Exploratory Technique for Investigating Large Quantities of Categorical Data , 1980 .

[9]  A. Tversky Elimination by aspects: A theory of choice. , 1972 .

[10]  Satoshi Fujii,et al.  Anticipated Travel Time, Information Acquisition, and Actual Experience: Hanshin Expressway Route Closure, Osaka-Sakai, Japan , 2000 .

[11]  Hjp Harry Timmermans,et al.  Inductive Learning Approach to Evolutionary Decision Processes in Activity-Scheduling Behavior: Theory and Numerical Experiments , 2001 .

[12]  Kay W. Axhausen,et al.  ADAPTING TO CHANGE: SOME EVIDENCE FROM A SIMPLE LEARNING MODEL , 1995 .

[13]  P.H.J. van der Mede,et al.  The impact of traffic information: Dynamics in route and departure time choice , 1993 .

[14]  J. Horowitz The stability of stochastic equilibrium in a two-link transportation network , 1984 .

[15]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[16]  Ta Theo Arentze,et al.  Modeling Learning and Evolutionary Adaptation Processes in Activity Settings: Theory and Numerical Simulations , 2000 .

[17]  J W Polak THE INFLUENCE OF ALTERNATIVE TRAVELLER LEARNING MECHANISMS ON THE DYNAMICS OF TRANSPORT SYSTEMS , 1998 .