TY - JOUR
T1 - Beyond traditional kernels: Classification in two dissimilarity-based representation spaces
AU - Pekalska, Elzbieta
AU - Duin, Robert P W
N1 - Pekalska, Elzbieta Duin, Robert P. W. British Engineering and Physical Science Research Council (EPSRC) [EP/D066883/1]; Dutch Organization for Scientific Research (NWO) This work was supported in part by the British Engineering and Physical Science Research Council (EPSRC) under Project EP/D066883/1 and in part by the Dutch Organization for Scientific Research (NWO). This paper was recommended by Associated Editor M. Last. 80 IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC PISCATAWAY 372KA
PY - 2008
Y1 - 2008
N2 - Proximity captures the degree of similarity between examples and is thereby fundamental in learning. Learning from pairwise proximity data usually relies on either kernel methods for specifically designed kernels or the nearest neighbor (NN) rule. Kernel methods are powerful, but often cannot handle arbitrary proximities without necessary corrections. The NN rule can work well in such cases, but suffers from local decisions. The aim of this paper is to provide an indispensable explanation and insights about two simple yet powerful alternatives when neither conventional kernel methods nor the NN rule can perform best. These strategies use two proximity-based representation spaces (RSs) in which accurate classifiers are trained on all training objects and demand comparisons to a small set of prototypes. They can handle all meaningful dissimilarity measures, including non-Euclidean and nonmetric ones. Practical examples illustrate that these RSs can be highly advantageous in supervised learning. Simple classifiers built there tend to outperform the NN rule. Moreover, computational complexity may be controlled. Consequently, these approaches offer an appealing alternative to learn from proximity data for which kernel methods cannot directly be applied, are too costly or impractical, while the NN rule leads to noisy results. © 2008 IEEE.
AB - Proximity captures the degree of similarity between examples and is thereby fundamental in learning. Learning from pairwise proximity data usually relies on either kernel methods for specifically designed kernels or the nearest neighbor (NN) rule. Kernel methods are powerful, but often cannot handle arbitrary proximities without necessary corrections. The NN rule can work well in such cases, but suffers from local decisions. The aim of this paper is to provide an indispensable explanation and insights about two simple yet powerful alternatives when neither conventional kernel methods nor the NN rule can perform best. These strategies use two proximity-based representation spaces (RSs) in which accurate classifiers are trained on all training objects and demand comparisons to a small set of prototypes. They can handle all meaningful dissimilarity measures, including non-Euclidean and nonmetric ones. Practical examples illustrate that these RSs can be highly advantageous in supervised learning. Simple classifiers built there tend to outperform the NN rule. Moreover, computational complexity may be controlled. Consequently, these approaches offer an appealing alternative to learn from proximity data for which kernel methods cannot directly be applied, are too costly or impractical, while the NN rule leads to noisy results. © 2008 IEEE.
KW - Classifier design and evaluation
KW - Indefinite kernels
KW - Similarity measures
KW - Statistical learning
U2 - 10.1109/TSMCC.2008.2001687
DO - 10.1109/TSMCC.2008.2001687
M3 - Article
VL - 38
SP - 729
EP - 744
JO - IEEE Transactions on Systems, Man and Cybernetics. Part C: Applications and Reviews
JF - IEEE Transactions on Systems, Man and Cybernetics. Part C: Applications and Reviews
IS - 6
ER -