Jamming mitigation in cognitive radio networks using a modified Q-learning algorithm

The jamming attack is one of the most severe threats in cognitive radio networks, because it can lead to network degradation and even denial of service. However, a cognitive radio can exploit its ability of dynamic spectrum access and its learning capabilities to avoid jammed channels. In this paper, we study how Q-learning can be used to learn the jammer strategy in order to pro-actively avoid jammed channels. The problem with Q-learning is that it needs a long training period to learn the behavior of the jammer. To address the above concern, we take advantage of the wideband spectrum sensing capabilities of the cognitive radio to speed up the learning process and we make advantage of the already learned information to minimize the number of collisions with the jammer during training. The effectiveness of this modified algorithm is evaluated by simulations in the presence of different jamming strategies and the simulation results are compared to the original Q-learning algorithm applied to the same scenarios.

[1]  Wajdi Alhakami,et al.  Spectrum sharing security and attacks in CRNs: a review , 2014 .

[2]  Michele Zorzi,et al.  JENNA: A Jamming Evasive Network-Coding Neighbor-Discovery Algorithm for Cognitive Radio Networks , 2010, 2010 IEEE International Conference on Communications Workshops.

[3]  Shamik Sengupta,et al.  Stochastic Model for Cognitive Radio Networks under Jamming Attacks and Honeypot-Based Prevention , 2014, ICDCN.

[4]  Joseph Mitola,et al.  Cognitive radio: making software radios more personal , 1999, IEEE Wirel. Commun..

[5]  C. Ribeiro A Tutorial on Reinforcement Learning Techniques , 1999 .

[6]  Chunsheng Xin,et al.  A game-theoretical anti-jamming scheme for cognitive radio networks , 2013, IEEE Network.

[7]  K. J. Ray Liu,et al.  Optimal Defense against Jamming Attacks in Cognitive Radio Networks Using the Markov Decision Process Approach , 2010, 2010 IEEE Global Telecommunications Conference GLOBECOM 2010.

[8]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[9]  Vivek S. Borkar,et al.  Stochastic Approximation for Nonexpansive Maps: Application to Q-Learning Algorithms , 1997, SIAM J. Control. Optim..

[10]  Csaba Szepesv Ari,et al.  Generalized Markov Decision Processes: Dynamic-programming and Reinforcement-learning Algorithms , 1996 .

[11]  Qusay H. Mahmoud,et al.  Cognitive Networks: Towards Self-Aware Networks , 2007 .

[12]  Yishay Mansour,et al.  Learning Rates for Q-learning , 2004, J. Mach. Learn. Res..

[13]  Roberto Di Pietro,et al.  Jamming mitigation in cognitive radio networks , 2013, IEEE Network.

[14]  Victor Balogun Anti-jamming Performance of Hybrid FEC code in the Presence of CRN Random Jammers , 2014 .

[15]  Ben J. A. Kröse,et al.  Learning from delayed rewards , 1995, Robotics Auton. Syst..

[16]  Gerald Tesauro,et al.  Extending Q-Learning to General Adaptive Multi-Agent Systems , 2003, NIPS.

[17]  Mainak Chatterjee,et al.  Collaborative jamming and collaborative defense in cognitive radio networks , 2013, Pervasive Mob. Comput..

[18]  Csaba Szepesvári,et al.  Generalized Markov Decision Processes: Dynamic-programming and Reinforcement-learning Algorithms , 2010 .