CUED at ProbSum 2023: Hierarchical Ensemble of Summarization Models
暂无分享,去创建一个
[1] M. Churpek,et al. Overview of the Problem List Summarization (ProbSum) 2023 Shared Task on Summarizing Patients’ Active Diagnoses and Problems from Electronic Health Record Progress Notes , 2023, BIONLP.
[2] Micah J. Smith,et al. Do We Still Need Clinical Language Models? , 2023, CHIL.
[3] Xi Victoria Lin,et al. OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization , 2022, ArXiv.
[4] Shenmin Zhang,et al. BioGPT: Generative Pre-trained Transformer for Biomedical Text Generation and Mining , 2022, Briefings Bioinform..
[5] M. Churpek,et al. Summarizing Patients’ Problems from Hospital Progress Notes Using Pre-trained Sequence-to-Sequence Models , 2022, COLING.
[6] Stella Rose Biderman,et al. GPT-NeoX-20B: An Open-Source Autoregressive Language Model , 2022, BIGSCIENCE.
[7] Ari S. Morcos,et al. Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time , 2022, ICML.
[8] David Grangier,et al. High Quality Rather than High Model Probability: Minimum Bayes Risk Decoding with Neural Metrics , 2021, Transactions of the Association for Computational Linguistics.
[9] Mark J. F. Gales,et al. Long-Span Summarization via Local Attention and Content Selection , 2021, ACL.
[10] Mark J. F. Gales,et al. Uncertainty Estimation in Autoregressive Structured Prediction , 2021, ICLR.
[11] Ahmed A. Rafea,et al. Automatic text summarization: A comprehensive survey , 2021, Expert Syst. Appl..
[12] M. Gales,et al. CUED_SPEECH at TREC 2020 Podcast Summarisation Track , 2020, TREC.
[13] Yassir Fathullah,et al. Ensemble Distillation Approaches for Grammatical Error Correction , 2020, ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[14] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[15] Doug Downey,et al. Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks , 2020, ACL.
[16] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[17] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[18] Balaji Lakshminarayanan,et al. Deep Ensembles: A Loss Landscape Perspective , 2019, ArXiv.
[19] Kilian Q. Weinberger,et al. BERTScore: Evaluating Text Generation with BERT , 2019, ICLR.
[20] Rajesh Ranganath,et al. ClinicalBERT: Modeling Clinical Notes and Predicting Hospital Readmission , 2019, ArXiv.
[21] Jaewoo Kang,et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining , 2019, Bioinform..
[22] Richard Socher,et al. A Deep Reinforced Model for Abstractive Summarization , 2017, ICLR.
[23] Markus Freitag,et al. Ensemble Distillation for Neural Machine Translation , 2017, ArXiv.
[24] Charles Blundell,et al. Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.
[25] Bowen Zhou,et al. SummaRuNNer: A Recurrent Neural Network Based Sequence Model for Extractive Summarization of Documents , 2016, AAAI.
[26] Rico Sennrich,et al. Improving Neural Machine Translation Models with Monolingual Data , 2015, ACL.
[27] Richard M. Schwartz,et al. Improved Word-Level System Combination for Machine Translation , 2007, ACL.
[28] Hichem Sahbi,et al. Consensus Network Decoding for Statistical Machine Translation System Combination , 2007, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07.
[29] Richard M. Schwartz,et al. Combining Outputs from Multiple Machine Translation Systems , 2007, NAACL.
[30] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[31] Shankar Kumar,et al. Minimum Bayes-Risk Decoding for Statistical Machine Translation , 2004, NAACL.
[32] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[33] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[34] Antal van den Bosch,et al. Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics , 2007 .