暂无分享,去创建一个
[1] Seungwoo Kang,et al. NewsCube: delivering multiple aspects of news to mitigate media bias , 2009, CHI.
[2] Daniel Jurafsky,et al. Linguistic Models for Analyzing and Detecting Biased Language , 2013, ACL.
[3] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[4] Craig Sanders,et al. Media Effects and Society , 2001 .
[5] Chunting Zhou,et al. Detecting Hallucinated Content in Conditional Neural Sequence Generation , 2021, FINDINGS.
[6] Preslav Nakov,et al. Tanbih: Get To Know What You Are Reading , 2019, EMNLP.
[7] August E. Grant,et al. The Spiral of Opportunity and Frame Resonance: Mapping the Issue Cycle in News and Public Discourse , 2001 .
[8] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[9] Yejin Choi,et al. PowerTransformer: Unsupervised Controllable Revision for Biased Language Correction , 2020, EMNLP.
[10] Taylor W. Brown,et al. Exposure to opposing views on social media can increase political polarization , 2018, Proceedings of the National Academy of Sciences.
[11] Mirella Lapata,et al. Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization , 2018, EMNLP.
[12] Junehwa Song,et al. Contrasting Opposing Views of News Articles on Contentious Issues , 2011, ACL.
[13] Riccardo Puglisi,et al. Being The New York Times: the Political Behaviour of a Newspaper , 2006 .
[14] Martha Larson,et al. Truth or Error? Towards systematic analysis of factual errors in abstractive summaries , 2020, EVAL4NLP.
[15] Noah A. Smith,et al. Shedding (a Thousand Points of) Light on Biased Language , 2010, Mturk@HLT-NAACL.
[16] Diyi Yang,et al. Automatically Neutralizing Subjective Bias in Text , 2019, AAAI.
[17] M. Reynolds,et al. HOW THE NEWS SHAPES OUR CIVIC AGENDA , 2009 .
[18] Pascale Fung,et al. Team yeon-zi at SemEval-2019 Task 4: Hyperpartisan News Detection by De-noising Weakly-labeled Data , 2019, SemEval@NAACL-HLT.
[19] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[20] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.