Classification from Triplet Comparison Data

Learning from triplet comparison data has been extensively studied in the context of metric learning, where we want to learn a distance metric between two instances, and ordinal embedding, where we want to learn an embedding in a Euclidean space of the given instances that preserve the comparison order as much as possible. Unlike fully labeled data, triplet comparison data can be collected in a more accurate and human-friendly way. Although learning from triplet comparison data has been considered in many applications, an important fundamental question of whether we can learn a classifier only from triplet comparison data without all the labels has remained unanswered. In this letter, we give a positive answer to this important question by proposing an unbiased estimator for the classification risk under the empirical risk minimization framework. Since the proposed method is based on the empirical risk minimization framework, it inherently has the advantage that any surrogate loss function and any model, including neural networks, can be easily applied. Furthermore, we theoretically establish an estimation error bound for the proposed empirical risk minimizer. Finally, we provide experimental results to show that our method empirically works well and outperforms various baseline methods.

[1]  Masashi Sugiyama,et al.  Learning Under Non-stationarity: Covariate Shift Adaptation by Importance Weighting , 2012 .

[2]  Ulrike von Luxburg,et al.  Lens Depth Function and k-Relative Neighborhood Graph: Versatile Tools for Ordinal Data Analysis , 2016, J. Mach. Learn. Res..

[3]  Inderjit S. Dhillon,et al.  Information-theoretic metric learning , 2006, ICML '07.

[4]  Zhi-Hua Zhou,et al.  A brief introduction to weakly supervised learning , 2018 .

[5]  Gang Niu,et al.  On the Minimal Supervision for Training Any Binary Classifier from Only Unlabeled Data , 2018, ICLR.

[6]  Gang Niu,et al.  Theoretical Comparisons of Positive-Unlabeled Learning against Positive-Negative Learning , 2016, NIPS.

[7]  Dacheng Tao,et al.  Correcting the Triplet Selection Bias for Triplet Loss , 2018, ECCV.

[8]  Ulrike von Luxburg,et al.  Boosting for Comparison-Based Learning , 2018, IJCAI.

[9]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[10]  Kilian Q. Weinberger,et al.  Stochastic triplet embedding , 2012, 2012 IEEE International Workshop on Machine Learning for Signal Processing.

[11]  L. Deng,et al.  The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web] , 2012, IEEE Signal Processing Magazine.

[12]  James Philbin,et al.  FaceNet: A unified embedding for face recognition and clustering , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[13]  Kui Wu,et al.  Sensor localization with Ring Overlapping based on Comparison of Received Signal Strength Indicator , 2004, 2004 IEEE International Conference on Mobile Ad-hoc and Sensor Systems (IEEE Cat. No.04EX975).

[14]  J. MacQueen Some methods for classification and analysis of multivariate observations , 1967 .

[15]  Gang Niu,et al.  Analysis of Learning from Positive and Unlabeled Data , 2014, NIPS.

[16]  Eric Heim,et al.  Efficiently and Effectively Learning Models of Similarity from Human Feedback , 2016 .

[17]  Antti Ukkonen,et al.  Relative distance comparisons with confidence judgements , 2019, SDM.

[18]  Javed A. Aslam,et al.  Scaling Up Ordinal Embedding: A Landmark Approach , 2019, ICML.

[19]  Vladimir Vapnik,et al.  The Nature of Statistical Learning , 1995 .

[20]  Gordon D. A. Brown,et al.  Absolute identification by relative judgment. , 2005, Psychological review.

[21]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[22]  Masashi Sugiyama,et al.  Classification From Pairwise Similarities/Dissimilarities and Unlabeled Data via Empirical Risk Minimization , 2019, Neural Computation.

[23]  Gang Niu,et al.  Classification from Pairwise Similarity and Unlabeled Data , 2018, ICML.

[24]  Ulrike von Luxburg,et al.  Kernel functions based on triplet comparisons , 2016, NIPS.

[25]  David J. Kriegman,et al.  Generalized Non-metric Multidimensional Scaling , 2007, AISTATS.

[26]  Matthäus Kleindessner,et al.  Machine learning in a setting of ordinal distance information , 2017 .

[27]  Thorsten Joachims,et al.  Learning a Distance Metric from Relative Comparisons , 2003, NIPS.

[28]  Y. Takane,et al.  Generalized Inverse Matrices , 2011 .

[29]  Ulrike von Luxburg,et al.  Comparison-Based Nearest Neighbor Search , 2017, AISTATS.

[30]  Harikrishna Narasimhan,et al.  On the Relationship Between Binary Classification, Bipartite Ranking, and Binary Class Probability Estimation , 2013, NIPS.

[31]  Roland Vollgraf,et al.  Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.

[32]  Gang Niu,et al.  Information-Theoretic Semi-Supervised Metric Learning via Entropy Regularization , 2012, Neural Computation.

[33]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[34]  Luping Shi,et al.  CIFAR10-DVS: An Event-Stream Dataset for Object Classification , 2017, Front. Neurosci..

[35]  Ulrike von Luxburg,et al.  Comparison-Based Random Forests , 2018, ICML.

[36]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[37]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .