Empirical Evaluation of Sequence-to-Sequence Models for Word Discovery in Low-resource Settings
暂无分享,去创建一个
[1] Charles Yang,et al. Recession Segmentation: Simpler Online Word Segmentation Using Limited Resources , 2010, CoNLL.
[2] Lori Lamel,et al. Dropping of the Class-Prefix Consonant, Vowel Elision and Automatic Phonological Mining in Embosi (Bantu C 25) , 2015 .
[3] Mark Johnson,et al. Why is English so easy to segment? , 2013, CMCL.
[4] T. Griffiths,et al. A Bayesian framework for word segmentation: Exploring the effects of context , 2009, Cognition.
[5] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[6] Sebastian Stüker,et al. A Very Low Resource Language Speech Corpus for Computational Language Documentation Experiments , 2017, LREC.
[7] David Chiang,et al. A case study on using speech-to-translation alignments for language documentation , 2017, ArXiv.
[8] Fedor Moiseev,et al. Analyzing Multi-Head Self-Attention: Specialized Heads Do the Heavy Lifting, the Rest Can Be Pruned , 2019, ACL.
[9] Peter Austin,et al. The Cambridge handbook of endangered languages , 2011 .
[10] Jian Li,et al. Multi-Head Attention with Disagreement Regularization , 2018, EMNLP.
[11] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[12] Laurent Besacier,et al. Pervasive Attention: 2D Convolutional Neural Networks for Sequence-to-Sequence Prediction , 2018, CoNLL.
[13] Sanjeev Khudanpur,et al. Librispeech: An ASR corpus based on public domain audio books , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[14] Ali Can Kocabiyikoglu,et al. Augmenting Librispeech with French Translations: A Multimodal Corpus for Direct Speech Translation Evaluation , 2018, LREC.
[15] Sebastian Stüker,et al. Breaking the Unwritten Language Barrier: The BULB Project , 2016, SLTU.
[16] Pierre Godard. Unsupervised word discovery for computational language documentation , 2019 .
[17] John DeNero,et al. Adding Interpretable Attention to Neural Translation Models Improves Word Alignment , 2019, ArXiv.
[18] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[19] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[20] Kaitao Song,et al. Hybrid Self-Attention Network for Machine Translation , 2018, ArXiv.
[21] Bowen Zhou,et al. A Structured Self-attentive Sentence Embedding , 2017, ICLR.
[22] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[23] Olivier Pietquin,et al. Listen and Translate: A Proof of Concept for End-to-End Speech-to-Text Translation , 2016, NIPS 2016.
[24] Lukás Burget,et al. Variational Inference for Acoustic Unit Discovery , 2016, Workshop on Spoken Language Technologies for Under-resourced Languages.
[25] Bowen Zhou,et al. TOWARDS SPEECH TRANSLATION OF NON WRITTEN LANGUAGES , 2006, 2006 IEEE Spoken Language Technology Workshop.
[26] Wen Wang,et al. Toward human-assisted lexical unit discovery without text resources , 2016, 2016 IEEE Spoken Language Technology Workshop (SLT).
[27] Aline Villavicencio,et al. Unsupervised Word Segmentation from Speech with Attention , 2018, INTERSPEECH.