Distributed self‐organizing bandwidth allocation for priority‐based bus communication

The raising complexity in distributed embedded systems makes it necessary that the communication of such systems organizes itself automatically. In this paper, we tackle the problem of sharing bandwidth on priority‐based buses. Based on a game theoretic model, reinforcement learning algorithms are proposed that use simple local rules to establish bandwidth sharing. The algorithms require little computational effort and no additional communication. Extensive experiments show that the proposed algorithms establish the desired properties without global knowledge ortextita priori information. It is proven that communication nodes using these algorithms can co‐exist with nodes using other scheduling techniques. Finally, we propose a procedure that helps to set the learning parameters according to the desired behavior. Copyright © 2011 John Wiley & Sons, Ltd.

[1]  Peter Dayan,et al.  Technical Note: Q-Learning , 2004, Machine Learning.

[2]  Jürgen Teich,et al.  CAN+: A new backward-compatible Controller Area Network (CAN) protocol with up to 16× higher data rates. , 2009, 2009 Design, Automation & Test in Europe Conference & Exhibition.

[3]  Giuseppe Lipari,et al.  Elastic Scheduling for Flexible Workload Management , 2002, IEEE Trans. Computers.

[4]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[5]  Michael L. Littman,et al.  Value-function reinforcement learning in Markov games , 2001, Cognitive Systems Research.

[6]  Ben E. Cline Tuning Q-Learning Parameters with a Genetic Algorithm , 2004 .

[7]  Jürgen Teich,et al.  A Self-Organizing Distributed Reinforcement Learning Algorithm to Achieve Fair Bandwidth Allocation for Priority-Based Bus Communication , 2010, 2010 13th IEEE International Symposium on Object/Component/Service-Oriented Real-Time Distributed Computing Workshops.

[8]  Paul Davidsson,et al.  Quantifying the Impact of Learning Algorithm Parameter Tuning , 2006, AAAI.

[9]  Tarak Gandhi,et al.  Vehicle Surround Capture: Survey of Techniques and a Novel Omni-Video-Based Approach for Dynamic Panoramic Surround Maps , 2006, IEEE Transactions on Intelligent Transportation Systems.

[10]  Ratan K. Guha,et al.  Fair bandwidth sharing in distributed systems: a game-theoretic approach , 2005, IEEE Transactions on Computers.

[11]  Martin Lauer,et al.  An Algorithm for Distributed Reinforcement Learning in Cooperative Multi-Agent Systems , 2000, ICML.

[12]  Alan Burns,et al.  Controller Area Network (CAN) schedulability analysis: Refuted, revisited and revised , 2007, Real-Time Systems.

[13]  D. Fudenberg,et al.  The Theory of Learning in Games , 1998 .

[14]  Gianluca Cena,et al.  A distributed mechanism to improve fairness in CAN networks , 1995, Proceedings 1995 IEEE International Workshop on Factory Communication Systems. WFCS'95.

[15]  Jürgen Teich,et al.  Self-organizing Bandwidth Sharing in Priority-Based Medium Access , 2009, 2009 Third IEEE International Conference on Self-Adaptive and Self-Organizing Systems.

[16]  Hermann Kopetz,et al.  TTP - A time-triggered protocol for fault-tolerant real-time systems , 1993, FTCS-23 The Twenty-Third International Symposium on Fault-Tolerant Computing.

[17]  Lucia Lo Bello,et al.  Automotive communications-past, current and future , 2005, 2005 IEEE Conference on Emerging Technologies and Factory Automation.

[18]  Bart De Schutter,et al.  A Comprehensive Survey of Multiagent Reinforcement Learning , 2008, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[19]  Alan Burns,et al.  Real Time Scheduling Theory: A Historical Perspective , 2004, Real-Time Systems.