Charmanteau: Character Embedding Models For Portmanteau Creation

Portmanteaus are a word formation phenomenon where two words are combined to form a new word. We propose character-level neural sequence-to-sequence (S2S) methods for the task of portmanteau generation that are end-to-end-trainable, language independent, and do not explicitly use additional phonetic information. We propose a noisy-channel-style model, which allows for the incorporation of unsupervised word lists, improving performance over a standard source-to-target model. This model is made possible by an exhaustive candidate generation strategy specifically enabled by the features of the portmanteau task. Experiments find our approach superior to a state-of-the-art FST-based baseline with respect to ground truth accuracy and human evaluation.

[1]  C. Strapparava,et al.  A Computational Approach to the Automation of Creative Naming , 2012, ACL.

[2]  Lei Yu,et al.  The Neural Noisy Channel , 2016, ICLR.

[3]  Outi Bat-El Selecting the best of the worst: the grammar of Hebrew blends , 1996, Phonology.

[4]  J. Algeo Blends, a Structural and Systemic View , 1977 .

[5]  Yulia Tsvetkov,et al.  Morphological Inflection Generation Using Character Sequence to Sequence Learning , 2015, NAACL.

[6]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[7]  Philipp Koehn,et al.  Statistical Significance Tests for Machine Translation Evaluation , 2004, EMNLP.

[8]  Harsh Jhamtani,et al.  Generating Appealing Brand Names , 2017, CICLing.

[9]  Fabian Monrose,et al.  Emergent faithfulness to morphological and semantic heads in lexical blends , 2014 .

[10]  Kevin Knight,et al.  Multi-Source Neural Translation , 2016, NAACL.

[11]  Carlos-Eduardo Piñeros,et al.  The creation of portmanteaus in the extragrammatical morphology of Spanish , 2004 .

[12]  Kevin Knight,et al.  How to Make a Frenemy: Multitape FSTs for Portmanteau Generation , 2015, HLT-NAACL.

[13]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[14]  Dan Ventura,et al.  Nehovah: A Neologism Creator Nomen Ipsum , 2014, ICCC.

[15]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[16]  Yoshua Bengio,et al.  A Character-level Decoder without Explicit Segmentation for Neural Machine Translation , 2016, ACL.

[17]  Gholamreza Haffari,et al.  Towards Decoding as Continuous Optimisation in Neural Machine Translation , 2017, EMNLP.

[18]  Paul Wexler,et al.  Studia linguistica et orientalia memoriae Haim Blanc dedicata , 1989 .

[19]  Gholamreza Haffari,et al.  Decoding as Continuous Optimization in Neural Machine Translation , 2017, ArXiv.

[20]  José A. R. Fonollosa,et al.  Character-based Neural Machine Translation , 2016, ACL.