Multi-Task Learning as a Bargaining Game

In Multi-task learning (MTL), a joint model is trained to simultaneously make predictions for several tasks. Joint training reduces computation costs and improves data efficiency; however, since the gradients of these different tasks may conflict, training a joint model for MTL often yields lower performance than its corresponding single-task counterparts. A common method for alleviating this issue is to combine per-task gradients into a joint update direction using a particular heuristic. In this paper, we propose viewing the gradients combination step as a bargaining game, where tasks negotiate to reach an agreement on a joint direction of parameter update. Under certain assumptions, the bargaining problem has a unique solution, known as the Nash Bargaining Solution , which we propose to use as a principled approach to multi-task learning. We describe a new MTL optimization procedure, Nash-MTL, and derive theoretical guarantees for its convergence. Empirically, we show that Nash-MTL achieves state-of-the-art results on multiple MTL benchmarks in various domains.

[1]  V. Farias,et al.  Fair Exploration via Axiomatic Bargaining , 2021, NeurIPS.

[2]  Joelle Pineau,et al.  Multi-Task Reinforcement Learning with Context-based Representations , 2021, ICML.

[3]  Yulia Tsvetkov,et al.  Gradient Vaccine: Investigating and Improving Multi-task Optimization in Massively Multilingual Models , 2020, ICLR.

[4]  Gal Chechik,et al.  Learning the Pareto Front with Hypernetworks , 2020, ICLR.

[5]  Ethan Fetaya,et al.  Auxiliary Learning by Implicit Differentiation , 2020, ICLR.

[6]  Gal Chechik,et al.  Self-Supervised Learning for Domain Adaptation on Point Clouds , 2020, 2021 IEEE Winter Conference on Applications of Computer Vision (WACV).

[7]  Yu Zhang,et al.  A Closer Look at Loss Weighting in Multi-Task Learning , 2021, ArXiv.

[8]  Sungwook Kim,et al.  Cooperative Federated Learning-Based Task Offloading Scheme for Tactical Edge Networks , 2021, IEEE Access.

[9]  I. Valera,et al.  Rotograd: Dynamic Gradient Homogenization for Multi-Task Learning , 2021, ArXiv.

[10]  Morteza Saberi,et al.  GBK-means clustering algorithm: An improvement to the K-means algorithm based on the bargaining game , 2021, Knowledge-Based Systems.

[11]  Dragomir Anguelov,et al.  Just Pick a Sign: Optimizing Deep Multitask Models with Gradient Sign Dropout , 2020, NeurIPS.

[12]  Michael Crawshaw,et al.  Multi-Task Learning with Deep Neural Networks: A Survey , 2020, ArXiv.

[13]  Yi Wu,et al.  Multi-Task Reinforcement Learning with Soft Modularization , 2020, NeurIPS.

[14]  Stephan Günnemann,et al.  Directional Message Passing for Molecular Graphs , 2020, ICLR.

[15]  S. Levine,et al.  Gradient Surgery for Multi-Task Learning , 2020, NeurIPS.

[16]  Jitendra Malik,et al.  Which Tasks Should Be Learned Together in Multi-task Learning? , 2019, ICML.

[17]  Patrick Taillandier,et al.  The Kalai-Smorodinsky solution for many-objective Bayesian optimization , 2019, J. Mach. Learn. Res..

[18]  S. Levine,et al.  Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning , 2019, CoRL.

[19]  Yike Guo,et al.  Regularizing Deep Multi-Task Networks using Orthogonal Gradients , 2019, ArXiv.

[20]  Yaron Lipman,et al.  Provably Powerful Graph Networks , 2019, NeurIPS.

[21]  Razvan Pascanu,et al.  Ray Interference: a Source of Plateaus in Deep Reinforcement Learning , 2019, ArXiv.

[22]  Iasonas Kokkinos,et al.  Attentive Single-Tasking of Multiple Tasks , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[23]  Jan Eric Lenssen,et al.  Fast Graph Representation Learning with PyTorch Geometric , 2019, ArXiv.

[24]  Andrew J. Davison,et al.  Self-Supervised Generalisation with Meta Auxiliary Learning , 2019, NeurIPS.

[25]  Andrew J. Davison,et al.  End-To-End Multi-Task Learning With Attention , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[26]  Michael I. Jordan,et al.  First-order methods almost always avoid saddle points: The case of vanishing step-sizes , 2019, NeurIPS.

[27]  Razvan Pascanu,et al.  Adapting Auxiliary Losses Using Gradient Similarity , 2018, ArXiv.

[28]  Ying Wu,et al.  A Modulation Module for Multi-task Learning with Applications in Image Retrieval , 2018, ECCV.

[29]  Chenguang Shi,et al.  Nash Bargaining Game-Theoretic Framework for Power Control in Distributed Multiple-Radar Architecture Underlying Wireless Communication System , 2018, Entropy.

[30]  Zhao Chen,et al.  GradNorm: Gradient Normalization for Adaptive Loss Balancing in Deep Multitask Networks , 2017, ICML.

[31]  Roberto Cipolla,et al.  Multi-task Learning Using Uncertainty to Weigh Losses for Scene Geometry and Semantics , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[32]  Qiang Yang,et al.  An Overview of Multi-task Learning , 2018 .

[33]  Sebastian Ruder,et al.  An Overview of Multi-Task Learning in Deep Neural Networks , 2017, ArXiv.

[34]  Samuel S. Schoenholz,et al.  Neural Message Passing for Quantum Chemistry , 2017, ICML.

[35]  Yoshimasa Tsuruoka,et al.  A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks , 2016, EMNLP.

[36]  Abhinav Gupta,et al.  Learning to push by grasping: Using multiple tasks for effective learning , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[37]  Roberto Cipolla,et al.  SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[38]  Stephen P. Boyd,et al.  Variations and extension of the convex–concave procedure , 2016 .

[39]  Sebastian Ramos,et al.  The Cityscapes Dataset for Semantic Urban Scene Understanding , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[40]  Jian Sun,et al.  Instance-Aware Semantic Segmentation via Multi-task Network Cascades , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[41]  Samy Bengio,et al.  Order Matters: Sequence to sequence for sets , 2015, ICLR.

[42]  Paul Strauss,et al.  Introduction To The Theory Of Games , 2016 .

[43]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[44]  H. Young,et al.  Handbook of Game Theory with Economic Applications , 2015 .

[45]  Xiaoou Tang,et al.  Facial Landmark Detection by Deep Multi-task Learning , 2014, ECCV.

[46]  Pavlo O. Dral,et al.  Quantum chemistry structures and properties of 134 kilo molecules , 2014, Scientific Data.

[47]  Martin Jaggi,et al.  Revisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization , 2013, ICML.

[48]  Derek Hoiem,et al.  Indoor Segmentation and Support Inference from RGBD Images , 2012, ECCV.

[49]  J. Désidéri Multiple-gradient descent algorithm (MGDA) for multiobjective optimization , 2012 .

[50]  Ephraim Zehavi,et al.  Smart carrier sensing for distributed computation of the generalized nash bargaining solution , 2011, 2011 17th International Conference on Digital Signal Processing (DSP).

[51]  Gert R. G. Lanckriet,et al.  On the Convergence of the Concave-Convex Procedure , 2009, NIPS.

[52]  Mohsen Guizani,et al.  A Cooperation Strategy Based on Nash Bargaining Solution in Cooperative Relay Networks , 2008, IEEE Transactions on Vehicular Technology.

[53]  Ferenc Szidarovszky,et al.  Multi-Agent Learning Model with Bargaining , 2006, Proceedings of the 2006 Winter Simulation Conference.

[54]  Rich Caruana,et al.  Multitask Learning , 1997, Machine Learning.

[55]  Alan L. Yuille,et al.  The Concave-Convex Procedure , 2003, Neural Computation.

[56]  Anthony T. Chronopoulos,et al.  Load balancing in distributed systems: an approach using cooperative games , 2002, Proceedings 16th International Parallel and Distributed Processing Symposium.

[57]  Jonathan Baxter,et al.  A Model of Inductive Bias Learning , 2000, J. Artif. Intell. Res..

[58]  N. Dagan,et al.  The bankruptcy problem: a cooperative bargaining approach , 1993 .

[59]  S. Hart,et al.  Handbook of Game Theory with Economic Applications , 1992 .

[60]  M. Bacharach Two-person Cooperative Games , 1976 .