Automatic classification of complaint letters according to service provider categories

Abstract In the technological age, the phenomenon of complaint letters published on the Internet is increasing. Therefore, it is important to automatically classify complaint letters according to various criteria, such as company categories. In this research, we investigated the automatic text classification of complaint letters written in Hebrew that were sent to various companies from a wide variety of categories. The classification was performed according to company categories such as insurance, cellular communication, and rental cars. We conducted an extensive set of classification experiments of complaint letters to seven/six/five/four company categories. The classification experiments were performed using various sets of word unigrams, four machine learning methods, two feature filtering methods, and parameter tuning. The classification results are relatively high for all six measures: accuracy, precision, recall, F1, PRC-area, and ROC-area. The best accuracy results for seven, six, five, and four categories are 84.5%, 88.4%, 91.4%, and 93.8%, respectively. An analysis of the most frequently occurring words in the complaints about almost all categories revealed that the most significant issues were related to poor service and delayed delivery. An interesting result shows that only in the domain of hospitals was the subject of the domain itself (i.e., the patient, the medical treatment, the place of the treatment, and the medical staff) the most important issue. Another interesting finding is that the issue of “price” was of little or no importance to the complainants. These findings suggest that in their preoccupation with their bottom line of profitability, many service providers are blind to how paramount good service and timely delivery (and, in the case of hospitals, the domain itself) are to their clientele.

[1]  Andrew P. Bradley,et al.  The use of the area under the ROC curve in the evaluation of machine learning algorithms , 1997, Pattern Recognit..

[2]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[3]  F. F. Reichheld,et al.  Zero defections: quality comes to services. , 1990, Harvard business review.

[4]  Mark Goadrich,et al.  The relationship between Precision-Recall and ROC curves , 2006, ICML.

[5]  Lillian Lee,et al.  Opinion Mining and Sentiment Analysis , 2008, Found. Trends Inf. Retr..

[6]  Johannes Fürnkranz,et al.  A Study Using $n$-gram Features for Text Categorization , 1998 .

[7]  Gavin L. Fox Getting Good Complaining without Bad Complaining , 2008 .

[8]  Mirna Adriani,et al.  Short Text Classification on Complaint Documents , 2016, Int. J. Comput. Linguistics Appl..

[9]  Philip H. Swain,et al.  Purdue e-Pubs , 2022 .

[10]  Yiannis Kompatsiaris,et al.  News Articles Classification Using Random Forests and Weighted Multimodal Features , 2014, IRFC.

[11]  Ohbyung Kwon,et al.  Effects of data set features on the performances of classification algorithms , 2013, Expert Syst. Appl..

[12]  Yu Xue,et al.  Text classification based on deep belief network and softmax regression , 2016, Neural Computing and Applications.

[13]  Yijun Li,et al.  Service Failure Complaints Identification in Social Media: A Text Classification Approach , 2013, ICIS.

[14]  Eibe Frank,et al.  Logistic Model Trees , 2003, Machine Learning.

[15]  Huan Liu,et al.  Unsupervised sentiment analysis with emotional signals , 2013, WWW.

[16]  Kevin Knight,et al.  Mining online text , 1999, Commun. ACM.

[17]  P Howell,et al.  Exchange of stuttering from function words to content words with age. , 1999, Journal of speech, language, and hearing research : JSLHR.

[18]  S. Sathiya Keerthi,et al.  Improvements to Platt's SMO Algorithm for SVM Classifier Design , 2001, Neural Computation.

[19]  S. Cessie,et al.  Ridge Estimators in Logistic Regression , 1992 .

[20]  Yaakov HaCohen-Kerner,et al.  Initial Experiments with Extraction of Stopwords in Hebrew , 2010, KDIR.

[21]  Thomas M. Tripp,et al.  When Unhappy Customers Strike Back on the Internet , 2011 .

[22]  Laurel J. Brinton,et al.  The Structure of Modern English: A linguistic introduction , 2000 .

[23]  Yaakov HaCohen-Kerner,et al.  WORDS AS CLASSIFIERS OF DOCUMENTS ACCORDING TO THEIR HISTORICAL PERIOD AND THE ETHNIC ORIGIN OF THEIR AUTHORS , 2008, Cybern. Syst..

[24]  Yaakov HaCohen-Kerner,et al.  Identifying the Correct Root of an Ambiguous Hebrew Word , 2014, Language, Culture, Computation.

[25]  Yaakov HaCohen-Kerner,et al.  AUTOMATIC MACHINE LEARNING OF KEYPHRASE EXTRACTION FROM SHORT HTML DOCUMENTS WRITTEN IN HEBREW , 2007, Cybern. Syst..

[26]  N. Prasath,et al.  Opinion mining and sentiment analysis on a Twitter data stream , 2012, International Conference on Advances in ICT for Emerging Regions (ICTer2012).

[27]  J Forster,et al.  A Cognitive Computing Approach for Classification of Complaints in the Insurance Industry , 2017 .

[28]  Takaya Saito,et al.  The Precision-Recall Plot Is More Informative than the ROC Plot When Evaluating Binary Classifiers on Imbalanced Datasets , 2015, PloS one.

[29]  Stephen S. Tax,et al.  Customer Evaluations of Service Complaint Experiences: Implications for Relationship Marketing: , 1998 .

[30]  Lukás Burget,et al.  Strategies for training large scale neural network language models , 2011, 2011 IEEE Workshop on Automatic Speech Recognition & Understanding.

[31]  George Forman,et al.  An Extensive Empirical Study of Feature Selection Metrics for Text Classification , 2003, J. Mach. Learn. Res..

[32]  Juan M. Corchado,et al.  Tokenising, Stemming and Stopword Removal on Anti-spam Filtering Domain , 2005, CAEPIA.

[33]  Efstathios Stamatatos,et al.  Words versus Character n-Grams for Anti-Spam Filtering , 2007, Int. J. Artif. Intell. Tools.

[34]  Yaakov HaCohen-Kerner,et al.  Stance Classification of Tweets Using Skip Char Ngrams , 2017, ECML/PKDD.

[35]  Shuly Wintner,et al.  Hebrew Computational Linguistics: Past and Future , 2004, Artificial Intelligence Review.

[36]  Pat Langley,et al.  Selection of Relevant Features and Examples in Machine Learning , 1997, Artif. Intell..

[37]  Mário J. Silva,et al.  Language identification in web pages , 2005, SAC '05.

[38]  Eibe Frank,et al.  Speeding Up Logistic Model Tree Induction , 2005, PKDD.

[39]  Tony J. Prescott,et al.  Analysing the Limitations of Deep Learning for Developmental Robotics , 2017, Living Machines.

[40]  Ian H. Witten,et al.  The WEKA data mining software: an update , 2009, SKDD.

[41]  B. Yersin,et al.  A retrospective study of complaint letters sent to a Swiss Emergency Department between 2009 and 2014. , 2018, Revue d'epidemiologie et de sante publique.

[42]  Wenqing Sun,et al.  Enhancing deep convolutional neural network scheme for breast cancer diagnosis with unlabeled data , 2017, Comput. Medical Imaging Graph..

[43]  HAADS: A Hebrew Aramaic abbreviation disambiguation system , 2010 .

[44]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[45]  J. Platt Sequential Minimal Optimization : A Fast Algorithm for Training Support Vector Machines , 1998 .

[46]  Kiyoharu Aizawa,et al.  Prediction of User Ratings of Oral Presentations using Label Relations , 2015, ASM@ACM Multimedia.

[47]  Andrew McCallum,et al.  A comparison of event models for naive bayes text classification , 1998, AAAI 1998.

[48]  Isti Surjandari Prajitno,et al.  Application of text mining for classification of textual reports: A study of Indonesia's national complaint handling system , 2016 .

[49]  Mary Caravella,et al.  Adding Bricks to Clicks: Predicting the Patterns of Cross-Channel Elasticities over Time , 2011 .

[50]  Yiming Yang,et al.  A Comparative Study on Feature Selection in Text Categorization , 1997, ICML.

[51]  Maite Taboada,et al.  Lexicon-Based Methods for Sentiment Analysis , 2011, CL.

[52]  Lloyd A. Smith,et al.  Practical feature subset selection for machine learning , 1998 .

[53]  Bini Omman,et al.  Evaluation of Features on Sentimental Analysis , 2015 .

[54]  Tom Fawcett,et al.  An introduction to ROC analysis , 2006, Pattern Recognit. Lett..

[55]  Yaakov HaCohen-Kerner,et al.  Summarization of Jewish Law Articles in Hebrew , 2003, CAINE.

[56]  Ron Kohavi,et al.  The Case against Accuracy Estimation for Comparing Induction Algorithms , 1998, ICML.

[57]  Shuhai Liu,et al.  A comparative study on text representation schemes in text categorization , 2005, Pattern Analysis and Applications.

[58]  Esther Levin,et al.  A recurrent neural network: Limitations and training , 1990, Neural Networks.

[59]  Efstathios Stamatatos,et al.  Learning to recognize webpage genres , 2009, Inf. Process. Manag..

[60]  John C. Haughey The Loyalty Effect , 1997, Business Ethics Quarterly.

[61]  Divakar Singh,et al.  A SURVEY REPORT ON TEXT CLASSIFICATION WITH DIFFERENT TERM WEIGHING METHODS AND COMPARISON BETWEEN CLASSIFICATION ALGORITHMS , 2013 .

[62]  Yaakov HaCohen-Kerner,et al.  Classifying True and False Hebrew Stories Using Word N-Grams , 2016, Cybern. Syst..

[63]  Paolo Rosso,et al.  On the impact of emotions on author profiling , 2016, Inf. Process. Manag..

[64]  Mark R. Lehto,et al.  Semiautomated Identification and Classification of Customer Complaints , 2013 .

[65]  Alexander Rossmann,et al.  Usage of Social Media Systems in Customer Service Strategies , 2017, HICSS.

[66]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[67]  Thorsten Joachims,et al.  Text Categorization with Support Vector Machines: Learning with Many Relevant Features , 1998, ECML.

[68]  Yaakov HaCohen-Kerner,et al.  Combined One Sense Disambiguation of Abbreviations , 2008, ACL.

[69]  J. Pennebaker,et al.  Psychological aspects of natural language. use: our words, our selves. , 2003, Annual review of psychology.

[70]  Fabrizio Sebastiani,et al.  Machine learning in automated text categorization , 2001, CSUR.

[71]  J. Hanley,et al.  The meaning and use of the area under a receiver operating characteristic (ROC) curve. , 1982, Radiology.

[72]  Daphne Koller,et al.  Toward Optimal Feature Selection , 1996, ICML.

[73]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques, 3rd Edition , 1999 .

[74]  J. Mayer Ensembles of Maximum Entropy , 1960 .

[75]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.