The Impact of Annotation Guidelines and Annotated Data on Extracting App Features from App Reviews

Annotation guidelines used to guide the annotation of training and evaluation datasets can have a considerable impact on the quality of machine learning models. In this study, we explore the effects of annotation guidelines on the quality of app feature extraction models. As a main result, we propose several changes to the existing annotation guidelines with a goal of making the extracted app features more useful and informative to the app developers. We test the proposed changes via simulating the application of the new annotation guidelines and then evaluating the performance of the supervised machine learning models trained on datasets annotated with initial and simulated guidelines. While the overall performance of automatic app feature extraction remains the same as compared to the model trained on the dataset with initial annotations, the features extracted by the model trained on the dataset with simulated new annotations are less noisy and more informative to the app developers. Secondly, we are interested in what kind of annotated training data is necessary for training an automatic app feature extraction model. In particular, we explore whether the training set should contain annotated app reviews from those apps/app categories on which the model is subsequently planned to be applied, or is it sufficient to have annotated app reviews from any app available for training, even when these apps are from very different categories compared to the test app. Our experiments show that having annotated training reviews from the test app is not necessary although including them into training set helps to improve recall. Furthermore, we test whether augmenting the training set with annotated product reviews helps to improve the performance of app feature extraction. We find that the models trained on augmented training set lead to improved recall but at the cost of the drop in precision.

[1]  Ning Chen,et al.  AR-miner: mining informative reviews for developers from mobile app marketplace , 2014, ICSE.

[2]  Zhiqiang Toh,et al.  DLIREC: Aspect Term Extraction and Term Polarity Classification System , 2014, *SEMEVAL.

[3]  Elhadi M. Shakshuki,et al.  Comparing mobile apps by identifying 'Hot' features , 2018, Future Gener. Comput. Syst..

[4]  Shafiq R. Joty,et al.  Fine-grained Opinion Mining with Recurrent Neural Networks and Word Embeddings , 2015, EMNLP.

[5]  Walid Maalej,et al.  How Do Users Like This Feature? A Fine Grained Sentiment Analysis of App Reviews , 2014, 2014 IEEE 22nd International Requirements Engineering Conference (RE).

[6]  Lina Zhou,et al.  RubE: Rule-based methods for extracting product features from online consumer reviews , 2017, Inf. Manag..

[7]  Peng Liang,et al.  Automatic Classification of Non-Functional Requirements from Augmented App User Reviews , 2017, EASE.

[8]  Walid Maalej,et al.  SAFE: A Simple Approach for Feature Extraction from App Descriptions and App Reviews , 2017, 2017 IEEE 25th International Requirements Engineering Conference (RE).

[9]  Tung Thanh Nguyen,et al.  Mining User Opinions in Mobile App Reviews: A Keyword-Based Approach (T) , 2015, 2015 30th IEEE/ACM International Conference on Automated Software Engineering (ASE).

[10]  Andrew McCallum,et al.  Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.

[11]  Bernd Brügge,et al.  Retrieving Diverse Opinions from App Reviews , 2015, 2015 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM).

[12]  Marcos André Gonçalves,et al.  A Feature-Oriented Sentiment Rating for Mobile App Reviews , 2018, WWW.

[13]  Bastin Tony Roy Savarimuthu,et al.  Approaches for prioritizing feature improvements extracted from app reviews , 2016, EASE.

[14]  Suresh Manandhar,et al.  SemEval-2014 Task 4: Aspect Based Sentiment Analysis , 2014, *SEMEVAL.

[15]  Rodrigo Agerri,et al.  EliXa: A Modular and Flexible ABSA Platform , 2015, *SEMEVAL.

[16]  Ulf Leser,et al.  SCARE ― The Sentiment Corpus of App Reviews with Fine-grained Annotations in German , 2016, LREC.

[17]  John Pavlopoulos,et al.  Aspect Term Extraction for Sentiment Analysis: New Datasets, New Evaluation Measures and an Improved Unsupervised Method , 2014 .

[18]  Chun Chen,et al.  Opinion Word Expansion and Target Extraction through Double Propagation , 2011, CL.

[19]  Jason Weston,et al.  Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..

[20]  Bing Liu,et al.  Aspect and Entity Extraction for Opinion Mining , 2014 .

[21]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[22]  Xiaodong Gu,et al.  "What Parts of Your Apps are Loved by Users?" (T) , 2015, 2015 30th IEEE/ACM International Conference on Automated Software Engineering (ASE).

[23]  Walid Maalej,et al.  User feedback in the appstore: An empirical study , 2013, 2013 21st IEEE International Requirements Engineering Conference (RE).

[24]  Michael I. Jordan,et al.  Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..

[25]  Harald C. Gall,et al.  How can i improve my app? Classifying user reviews for software maintenance and evolution , 2015, 2015 IEEE International Conference on Software Maintenance and Evolution (ICSME).

[26]  Dietmar Pfahl,et al.  Feature-based evaluation of competing apps , 2016, WAMA@SIGSOFT FSE.

[27]  Walid Maalej,et al.  Automatically Classifying Functional and Non-functional Requirements Using Supervised Machine Learning , 2017, 2017 IEEE 25th International Requirements Engineering Conference (RE).

[28]  Erik Cambria,et al.  Aspect extraction for opinion mining with a deep convolutional neural network , 2016, Knowl. Based Syst..

[29]  Maleknaz Nayebi,et al.  Toward Data-Driven Requirements Engineering , 2016, IEEE Software.