A generalization of Haussler's convolution kernel: mapping kernel

Haussler's convolution kernel provides a successful framework for engineering new positive semidefinite kernels, and has been applied to a wide range of data types and applications. In the framework, each data object represents a finite set of finer grained components. Then, Haussler's convolution kernel takes a pair of data objects as input, and returns the sum of the return values of the predetermined primitive positive semidefinite kernel calculated for all the possible pairs of the components of the input data objects. On the other hand, the mapping kernel that we introduce in this paper is a natural generalization of Haussler's convolution kernel, in that the input to the primitive kernel moves over a predetermined subset rather than the entire cross product. Although we have plural instances of the mapping kernel in the literature, their positive semidefiniteness was investigated in case-by-case manners, and worse yet, was sometimes incorrectly concluded. In fact, there exists a simple and easily checkable necessary and sufficient condition, which is generic in the sense that it enables us to investigate the positive semidefiniteness of an arbitrary instance of the mapping kernel. This is the first paper that presents and proves the validity of the condition. In addition, we introduce two important instances of the mapping kernel, which we refer to as the size-of-index-structure-distribution kernel and the editcost-distribution kernel. Both of them are naturally derived from well known (dis)similarity measurements in the literature (e.g. the maximum agreement tree, the edit distance), and are reasonably expected to improve the performance of the existing measures by evaluating their distributional features rather than their peak (maximum/minimum) features.

[1]  Kunihiko Sadakane,et al.  Rooted Maximum Agreement Supertrees , 2004, Algorithmica.

[2]  Nello Cristianini,et al.  An Introduction to Support Vector Machines and Other Kernel-based Learning Methods , 2000 .

[3]  Jason Weston,et al.  Mismatch string kernels for discriminative protein classification , 2004, Bioinform..

[4]  David Haussler,et al.  Convolution kernels on discrete structures , 1999 .

[5]  François Nicolas,et al.  Maximum Agreement and Compatible Supertrees (Extended Abstract) , 2004, CPM.

[6]  Nello Cristianini,et al.  Classification using String Kernels , 2000 .

[7]  Michael J. Fischer,et al.  The String-to-String Correction Problem , 1974, JACM.

[8]  Tao Jiang,et al.  A Class of Edit Kernels for SVMs to Predict Translation Initiation Sites in Eukaryotic mRNAs , 2005, J. Comput. Biol..

[9]  Marcello Pelillo,et al.  Matching Free Trees, Maximal Cliques, and Monotone Game Dynamics , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[10]  Chuan Yi Tang,et al.  A New Measure of Edit Distance between Labeled Trees , 2001, COCOON.

[11]  Tao Jiang,et al.  On the Complexity of Comparing Evolutionary Trees , 1996, Discret. Appl. Math..

[12]  Kaizhong Zhang,et al.  Algorithms for the constrained editing distance between ordered labeled trees and related problems , 1995, Pattern Recognit..

[13]  Paolo Frasconi,et al.  Weighted decomposition kernels , 2005, ICML.

[14]  Hisashi Kashima,et al.  Kernels for Semi-Structured Data , 2002, ICML.

[15]  Steven Salzberg,et al.  A method for identifying splice sites and translational start sites in eukaryotic mRNA , 1997, Comput. Appl. Biosci..

[16]  Tatsuya Akutsu,et al.  Efficient tree-matching methods for accurate carbohydrate database queries. , 2003, Genome informatics. International Conference on Genome Informatics.

[17]  Horst Bunke,et al.  On a relation between graph edit distance and maximum common subgraph , 1997, Pattern Recognit. Lett..

[18]  Michael Collins,et al.  Convolution Kernels for Natural Language , 2001, NIPS.

[19]  Kuo-Chung Tai,et al.  The Tree-to-Tree Correction Problem , 1979, JACM.

[20]  Jun Suzuki,et al.  Convolution Kernels with Feature Selection for Natural Language Processing Tasks , 2004, ACL.