Compositionality and Generalization In Emergent Languages

Natural language allows us to refer to novel composite concepts by combining expressions denoting their parts according to systematic rules, a property known as \emph{compositionality}. In this paper, we study whether the language emerging in deep multi-agent simulations possesses a similar ability to refer to novel primitive combinations, and whether it accomplishes this feat by strategies akin to human-language compositionality. Equipped with new ways to measure compositionality in emergent languages inspired by disentanglement in representation learning, we establish three main results. First, given sufficiently large input spaces, the emergent language will naturally develop the ability to refer to novel composite concepts. Second, there is no correlation between the degree of compositionality of an emergent language and its ability to generalize. Third, while compositionality is not necessary for generalization, it provides an advantage in terms of language transmission: The more compositional a language is, the more easily it will be picked up by new learners, even when the latter differ in architecture from the original agents. We conclude that compositionality does not arise from simple generalization pressure, but if an emergent language does chance upon it, it will be more likely to survive and thrive.

[1]  Pieter Abbeel,et al.  Emergence of Grounded Compositional Language in Multi-Agent Populations , 2017, AAAI.

[2]  Frank van der Velde,et al.  Lack of combinatorial productivity in language processing with simple recurrent networks , 2004, Connect. Sci..

[3]  S. Kirby,et al.  Compression and communication in the cultural evolution of linguistic structure , 2015, Cognition.

[4]  Steven T. Piantadosi,et al.  The communicative function of ambiguity in language , 2011, Cognition.

[5]  W. Strange Evolution of language. , 1984, JAMA.

[6]  Simon Kirby,et al.  Spontaneous evolution of linguistic structure-an iterated learning model of the emergence of regularity and irregularity , 2001, IEEE Trans. Evol. Comput..

[7]  Dennis Ulmer,et al.  On the Realization of Compositionality in Neural Networks , 2019, BlackboxNLP@ACL.

[8]  Martin A. Nowak,et al.  The evolution of syntactic communication , 2000, Nature.

[9]  D. Bickerton More Than Nature Needs: Language, Mind, and Evolution , 2014 .

[10]  Pieter Abbeel,et al.  Gradient Estimation Using Stochastic Computation Graphs , 2015, NIPS.

[11]  Andrew M. Dai,et al.  Capacity, Bandwidth, and Compositionality in Emergent Language Learning , 2020, AAMAS.

[12]  Jennifer G. Dy,et al.  Rate-Regularization and Generalization in VAEs , 2019 .

[13]  Eugene Kharitonov,et al.  Anti-efficient encoding in emergent communication , 2019, NeurIPS.

[14]  M. E. Medina-Callarotti Origins of Language , 2000 .

[15]  Philemon Brakel,et al.  Strong systematicity in sentence processing by simple recurrent networks , 2009 .

[16]  Dag Westerståhl,et al.  Compositionality II: Arguments and Problems , 2010 .

[17]  Michael Bowling,et al.  Ease-of-Teaching and Language Structure from Emergent Communication , 2019, NeurIPS.

[18]  Yoshua Bengio,et al.  Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.

[19]  N. Masataka The Origins of Language , 2008 .

[20]  Pascal Vincent,et al.  Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[21]  Simon Kirby,et al.  Sequence Memory Constraints Give Rise to Language-Like Structure through Iterated Learning , 2017, PloS one.

[22]  Mathijs Mul,et al.  The compositionality of neural networks: integrating symbolism and connectionism , 2019, ArXiv.

[23]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[24]  Marco Baroni,et al.  Generalization without Systematicity: On the Compositional Skills of Sequence-to-Sequence Recurrent Networks , 2017, ICML.

[25]  Simon Kirby,et al.  Iconicity and the Emergence of Combinatorial Structure in Language , 2016, Cogn. Sci..

[26]  Stefan Bauer,et al.  Robustly Disentangled Causal Mechanisms: Validating Deep Representations for Interventional Robustness , 2018, ICML.

[27]  Michael Cogswell,et al.  Emergence of Compositional Language with Deep Generational Transmission , 2019, ArXiv.

[28]  Eugene Kharitonov,et al.  EGG: a toolkit for research on Emergence of lanGuage in Games , 2019, EMNLP.

[29]  Barbara H. Partee,et al.  Compositionality in Formal Semantics , 2004 .

[30]  Roger B. Grosse,et al.  Isolating Sources of Disentanglement in Variational Autoencoders , 2018, NeurIPS.

[31]  Jacob Andreas,et al.  Measuring Compositionality in Representation Learning , 2019, ICLR.

[32]  Ben Poole,et al.  Categorical Reparameterization with Gumbel-Softmax , 2016, ICLR.

[33]  K. Zuberbühler,et al.  Compositionality in animals and humans , 2018, PLoS biology.

[34]  Mathijs Mul,et al.  Compositionality Decomposed: How do Neural Networks Generalise? , 2019, J. Artif. Intell. Res..

[35]  Ernest Lepore,et al.  The compositionality papers , 2002 .

[36]  Territoire Urbain,et al.  Convention , 1955, Hidden Nature.

[37]  Bernhard Schölkopf,et al.  Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations , 2018, ICML.

[38]  Eugene Kharitonov,et al.  Emergent Language Generalization and Acquisition Speed are not tied to Compositionality , 2020, BLACKBOXNLP.

[39]  Adele E. Goldberg,et al.  Explain Me This , 2019 .

[40]  Ronald J. Williams,et al.  Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.

[41]  Yi Ren,et al.  The Emergence of Compositional Languages for Numeric Concepts Through Iterated Learning in Neural Agents , 2019, ArXiv.

[42]  Christopher Burgess,et al.  beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.

[43]  Jennifer G. Dy,et al.  Evaluating Combinatorial Generalization in Variational Autoencoders , 2019, ArXiv.

[44]  Yee Whye Teh,et al.  The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables , 2016, ICLR.

[45]  Stefano Ermon,et al.  Bias and Generalization in Deep Generative Models: An Empirical Study , 2018, NeurIPS.

[46]  Stephen Clark,et al.  Emergence of Linguistic Communication from Referential Games with Symbolic and Pixel Input , 2018, ICLR.

[47]  Ryan M. Nefdt A Puzzle concerning Compositionality in Machines , 2020, Minds and Machines.

[48]  José M. F. Moura,et al.  Natural Language Does Not Emerge ‘Naturally’ in Multi-Agent Dialog , 2017, EMNLP.

[49]  Simon Kirby,et al.  Understanding Linguistic Evolution by Visualizing the Emergence of Topographic Mappings , 2006, Artificial Life.

[50]  Andriy Mnih,et al.  Disentangling by Factorising , 2018, ICML.