An Experimental Study on Learning with Good Edit Similarity Functions

Abstract : Similarity functions are essential to many learning algorithms. To allow their use in support vector machines (SVM), i.e., for the convergence of the learning algorithm to be guaranteed, they must be valid kernels. In the case of structured data, the similarities based on the popular edit distance often do not satisfy this requirement, which explains why they are typically used with k-nearest neighbor (k-NN). A common approach to use such edit similarities anyway in SVM is to transform them into potentially (but not provably) valid kernels. Recently, a different theory of learning with (epsilon,gamma,tau)-good similarity functions was proposed, allowing the use of non-kernel similarity functions. Moreover, the resulting models are supposedly sparse, as opposed to standard SVM models that can be unnecessarily dense. In this paper, we study the relevance and applicability of this theory in the context of string edit similarities. We show that they are naturally good for a given string classification task and provide experimental evidence that the obtained models not only clearly outperform the k-NN approach, but are also competitive with standard SVM models learned with state-of-the-art edit kernels, while being much sparser.
Document type :
Conference papers
Complete list of metadatas

https://hal.archives-ouvertes.fr/hal-00618706
Contributor : Marc Sebban <>
Submitted on : Friday, September 2, 2011 - 3:50:08 PM
Last modification on : Tuesday, September 10, 2019 - 11:32:08 AM

Identifiers

  • HAL Id : hal-00618706, version 1

Citation

Aurélien Bellet, Amaury Habrard, Marc Sebban. An Experimental Study on Learning with Good Edit Similarity Functions. ICTAI 2011, Nov 2011, United States. ⟨hal-00618706⟩

Share

Metrics

Record views

117