Wrapper positive Bayesian network classifiers

In the information retrieval framework, there are problems where the goal is to recover objects of a particular class from big sets of unlabelled objects. In some of these problems, only examples from the class we want to recover are available. For such problems, the machine learning community has developed algorithms that are able to learn binary classifiers in the absence of negative examples. Among them, we can find the positive Bayesian network classifiers, algorithms that induce Bayesian network classifiers from positive and unlabelled examples. The main drawback of these algorithms is that they require some previous knowledge about the a priori probability distribution of the class. In this paper, we propose a wrapper approach to tackle the learning when no such information is available, setting this probability at the optimal value in terms of the recovery of positive examples. The evaluation of classifiers in positive unlabelled learning problems is a non-trivial question. We have also worked on this problem, and we have proposed a new guiding metric to be used in the search for the optimal a priori probability of the positive class that we have called the pseudo F. We have empirically tested the proposed metric and the wrapper classifiers on both synthetic and real-life datasets. The results obtained in this empirical comparison show that the wrapper Bayesian network classifiers provide competitive results, particularly when the actual a priori probability of the positive class is high.

[1]  Ron Kohavi,et al.  Wrappers for Feature Subset Selection , 1997, Artif. Intell..

[2]  Nir Friedman,et al.  Bayesian Network Classifiers , 1997, Machine Learning.

[3]  Xing-Ming Zhao,et al.  Gene function prediction using labeled and unlabeled data , 2008, BMC Bioinformatics.

[4]  Charles Elkan,et al.  Learning gene regulatory networks from only positive and unlabeled data , 2010, BMC Bioinformatics.

[5]  Duoqian Miao,et al.  A Rough Set Approach to Classifying Web Page Without Negative Examples , 2007, PAKDD.

[6]  Frann Cois Denis,et al.  PAC Learning from Positive Statistical Queries , 1998, ALT.

[7]  Chris H. Q. Ding,et al.  PSoL: a positive sample only learning algorithm for finding non-coding RNA genes , 2006, Bioinform..

[8]  Y. Benjamini,et al.  THE CONTROL OF THE FALSE DISCOVERY RATE IN MULTIPLE TESTING UNDER DEPENDENCY , 2001 .

[9]  Fábio Gagliardi Cozman,et al.  Random Generation of Bayesian Networks , 2002, SBIA.

[10]  Xiaoli Li,et al.  Learning to Classify Texts Using Positive and Unlabeled Data , 2003, IJCAI.

[11]  Robert Castelo,et al.  Splice site identification by idlBNs , 2004, ISMB/ECCB.

[12]  Igor Kononenko,et al.  Semi-Naive Bayesian Classifier , 1991, EWSL.

[13]  Michael J. Pazzani,et al.  Searching for Dependencies in Bayesian Classifiers , 1995, AISTATS.

[14]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[15]  Xue Li,et al.  Dynamic classifier ensemble for positive unlabeled text stream classification , 2012, Knowledge and Information Systems.

[16]  J. A. Lozano,et al.  Prioritization of candidate cancer genes—an aid to oncogenomic studies , 2008, Nucleic acids research.

[17]  Pedro Larrañaga,et al.  A partially supervised classification approach to dominant and recessive human disease gene prediction , 2007, Comput. Methods Programs Biomed..

[18]  Zhoujun Li,et al.  Predicting gene function using few positive examples and unlabeled ones , 2010, BMC Genomics.

[19]  Hiroya Takamura,et al.  Cool Blog Classification from Positive and Unlabeled Examples , 2009, PAKDD.

[20]  Rémi Gilleron,et al.  Learning from positive and unlabeled examples , 2000, Theor. Comput. Sci..

[21]  Alexander Zien,et al.  Semi-Supervised Learning , 2006 .

[22]  José Antonio Lozano,et al.  Sensitivity Analysis of k-Fold Cross Validation in Prediction Error Estimation , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[23]  Marvin Minsky,et al.  Steps toward Artificial Intelligence , 1995, Proceedings of the IRE.

[24]  Jiawei Han,et al.  PEBL: Web page classification without negative examples , 2004, IEEE Transactions on Knowledge and Data Engineering.

[25]  Kevin Chen-Chuan Chang,et al.  PEBL: positive example based learning for Web page classification using SVM , 2002, KDD.

[26]  Borja Calvo Molinos Positive unlabelled learning with applications in computational biology , 2008 .

[27]  Bing Liu,et al.  Learning with Positive and Unlabeled Examples Using Weighted Logistic Regression , 2003, ICML.

[28]  Philip S. Yu,et al.  Building text classifiers using positive and unlabeled examples , 2003, Third IEEE International Conference on Data Mining.

[29]  See-Kiong Ng,et al.  Negative Training Data Can be Harmful to Text Classification , 2010, EMNLP.

[30]  Pedro Larrañaga,et al.  Feature subset selection from positive and unlabelled examples , 2009, Pattern Recognit. Lett..

[31]  Wen-Lian Hsu,et al.  Exploiting likely-positive and unlabeled data to improve the identification of protein-protein interaction articles , 2008, BMC Bioinformatics.

[32]  Pedro Larrañaga,et al.  Learning Bayesian classifiers from positive and unlabeled examples , 2007, Pattern Recognit. Lett..

[33]  Charles Elkan,et al.  Learning classifiers from only positive and unlabeled data , 2008, KDD.

[34]  Rémi Gilleron,et al.  Text Classification from Positive and Unlabeled Examples , 2002 .

[35]  Philip S. Yu,et al.  Partially Supervised Classification of Text Documents , 2002, ICML.

[36]  Louise C. Showe,et al.  Learning from positive examples when the negative class is undetermined- microRNA gene identification , 2008, Algorithms for Molecular Biology.

[37]  Philip S. Yu Editorial: State of the Transactions , 2004, IEEE Trans. Knowl. Data Eng..

[38]  Mark R. Segal,et al.  Biological sequence classification utilizing positive and unlabeled data , 2008, Bioinform..