Detecting flames and insults in text
暂无分享,去创建一个
While the Internet has become the leading source of information, it is also become the medium for flames, insults and other forms of abusive language, which add nothing to the quality of information available. A human reader can easily distinguish between what is information and what is a flame or any other form of abuse. It is however much more difficult for a language processor to do this automatically. This paper describes a new approach for an automated system to distinguish between information and personal attacks containing insulting or abusive expressions in a given document. In Linguistics, insulting or abusive messages are viewed as an extreme subset of the subjective language because of its extreme nature. We create a set of rules to extract the semantic information of a given sentence from the general semantic structure of that sentence to separate information from abusive language.
[1] Michael A. Covington,et al. A Fundamental Algorithm for Dependency Parsing , 2004 .
[2] Claire Cardie,et al. Annotating Expressions of Opinions and Emotions in Language , 2005, Lang. Resour. Evaluation.
[3] Janyce Wiebe,et al. Learning Subjective Language , 2004, CL.
[4] Ellen Spertus,et al. Smokey: Automatic Recognition of Hostile Messages , 1997, AAAI/IAAI.