Contingency-Aware Influence Maximization: A Reinforcement Learning Approach

The influence maximization (IM) problem aims at finding a subset of seed nodes in a social network that maximize the spread of influence. In this study, we focus on a sub-class of IM problems, where whether the nodes are willing to be the seeds when being invited is uncertain, called contingency-aware IM. Such contingency aware IM is critical for applications for non-profit organizations in low resource communities (e.g., spreading awareness of disease prevention). Despite the initial success, a major practical obstacle in promoting the solutions to more communities is the tremendous runtime of the greedy algorithms and the lack of high performance computing (HPC) for the non-profits in the field – whenever there is a new social network, the non-profits usually do not have the HPCs to recalculate the solutions. Motivated by this and inspired by the line of works that use reinforcement learning (RL) to address combinatorial optimization on graphs, we formalize the problem as a Markov Decision Process (MDP), and use RL to learn an IM policy over historically seen networks, and generalize to unseen networks with negligible runtime at test phase. To fully exploit the properties of our targeted problem, we propose two technical innovations that improve the existing methods, including state-abstraction and theoretically grounded reward shaping. Empirical results show that our method achieves influence as high as the state-of-the-art methods for contingency-aware IM, while having negligible runtime at test phase.

[1]  Éva Tardos,et al.  Maximizing the Spread of Influence through a Social Network , 2015, Theory Comput..

[2]  Max Welling,et al.  Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.

[3]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[4]  Shahin Jabbari,et al.  Active Screening for Recurrent Diseases: A Reinforcement Learning Approach , 2021, AAMAS.

[5]  Sergio Gomez Colmenarejo,et al.  Hybrid computing using a neural network with dynamic external memory , 2016, Nature.

[6]  Liwei Wang,et al.  Deep Reinforcement Learning-Based Approach to Tackle Topic-Aware Influence Maximization , 2020, Data Science and Engineering.

[7]  Jacob Goldenberg,et al.  Talk of the Network: A Complex Systems Look at the Underlying Process of Word-of-Mouth , 2001 .

[8]  Kai Han,et al.  Efficient Algorithms for Adaptive Influence Maximization , 2018, Proc. VLDB Endow..

[9]  Philip S. Yu,et al.  Multi-Round Influence Maximization , 2018, KDD.

[10]  Alexandre Lacoste,et al.  Learning Heuristics for the TSP by Policy Gradient , 2018, CPAIOR.

[11]  E. Rice,et al.  Examining Impacts of a Peer-Based Mindfulness and Yoga Intervention to Reduce Interpersonal Violence Among Young Adults Experiencing Homelessness , 2021, Journal of the Society for Social Work and Research.

[12]  Milind Tambe,et al.  Influence Maximization in Unknown Social Networks: Learning Policies for Effective Graph Sampling , 2020, AAMAS.

[13]  XiaoXiaokui,et al.  Efficient algorithms for adaptive influence maximization , 2018, VLDB 2018.

[14]  Ronald J. Williams,et al.  Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.

[15]  Andreas Krause,et al.  Cost-effective outbreak detection in networks , 2007, KDD '07.

[16]  Tavpritesh Sethi,et al.  Learning Explainable Interventions to Mitigate HIV Transmission in Sex Workers Across Five States in India , 2020, ArXiv.

[17]  Marc Esteve,et al.  Artificial intelligence for the public sector: opportunities and challenges of cross-sector collaboration , 2018, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[18]  Lawrence V. Snyder,et al.  Reinforcement Learning for Solving the Vehicle Routing Problem , 2018, NeurIPS.

[19]  Akash Mittal,et al.  GCOMB: Learning Budget-constrained Combinatorial Algorithms over Billion-sized Graphs , 2020, NeurIPS.

[20]  A-L Barabási,et al.  Structure and tie strengths in mobile communication networks , 2006, Proceedings of the National Academy of Sciences.

[21]  Xavier Bresson,et al.  An Efficient Graph Convolutional Network Technique for the Travelling Salesman Problem , 2019, ArXiv.

[22]  Wei Chen,et al.  Efficient approximation algorithms for adaptive influence maximization , 2020, The VLDB Journal.

[23]  Dhaya Sindhu Battina RESEARCH ON ARTIFICIAL INTELLIGENCE FOR CITIZEN SERVICES AND GOVERNMENT , 2021 .

[24]  Hongzi Mao,et al.  Learning scheduling algorithms for data processing clusters , 2018, SIGCOMM.

[25]  Jiangtao Cui,et al.  DISCO: Influence Maximization Meets Network Embedding and Deep Learning , 2019, ArXiv.

[26]  Samy Bengio,et al.  Neural Combinatorial Optimization with Reinforcement Learning , 2016, ICLR.

[27]  Zhuwen Li,et al.  Combinatorial Optimization with Graph Convolutional Networks and Guided Tree Search , 2018, NeurIPS.

[28]  Le Song,et al.  2 Common Formulation for Greedy Algorithms on Graphs , 2018 .

[29]  Le Song,et al.  Discriminative Embeddings of Latent Variable Models for Structured Data , 2016, ICML.

[30]  Jonathan Guo,et al.  The Application of Medical Artificial Intelligence Technology in Rural Areas of Developing Countries , 2018, Health equity.

[31]  Max Welling,et al.  Attention, Learn to Solve Routing Problems! , 2018, ICLR.

[32]  Christian Borgs,et al.  Maximizing Social Influence in Nearly Optimal Time , 2012, SODA.

[33]  Shou-De Lin,et al.  A Learning-based Framework to Handle Multi-round Multi-party Influence Maximization on Social Networks , 2015, KDD.

[34]  Andreas Krause,et al.  Adaptive Submodularity: Theory and Applications in Active Learning and Stochastic Optimization , 2010, J. Artif. Intell. Res..

[35]  Noseong Park,et al.  MONSTOR: An Inductive Approach for Estimating and Maximizing Influence over Unseen Social Networks , 2020, ArXiv.

[36]  Milind Tambe,et al.  End-to-End Influence Maximization in the Field , 2018, AAMAS.

[37]  Ben J. A. Kröse,et al.  Learning from delayed rewards , 1995, Robotics Auton. Syst..

[38]  Navdeep Jaitly,et al.  Pointer Networks , 2015, NIPS.

[39]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[40]  Haifeng Xu,et al.  Using Social Networks to Aid Homeless Shelters: Dynamic Influence Maximization under Uncertainty , 2016, AAMAS.

[41]  Bo An,et al.  Dynamic Electronic Toll Collection via Multi-Agent Deep Reinforcement Learning with Edge-Based Graph Convolutional Networks , 2019, IJCAI.

[42]  Leandro Soriano Marcolino,et al.  Please be an Influencer?: Contingency-Aware Influence Maximization , 2018, AAMAS.

[43]  Yi-Shin Chen,et al.  Boosting Reinforcement Learning in Competitive Influence Maximization with Transfer Learning , 2018, 2018 IEEE/WIC/ACM International Conference on Web Intelligence (WI).

[44]  Matthew Richardson,et al.  Mining the network value of customers , 2001, KDD '01.

[45]  Eric Rice,et al.  Clinical trial of an AI-augmented intervention for HIV prevention in youth experiencing homelessness , 2020, AAAI.

[46]  Xiaokui Xiao,et al.  Influence Maximization in Near-Linear Time: A Martingale Approach , 2015, SIGMOD Conference.

[47]  Yoshua Bengio,et al.  Machine Learning for Combinatorial Optimization: a Methodological Tour d'Horizon , 2018, Eur. J. Oper. Res..

[48]  Viktor K. Prasanna,et al.  Network-based intervention strategies to reduce violence among homeless , 2019, Social Network Analysis and Mining.

[49]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.