Adaptive Local Dissimilarity Measures for Discriminative Dimension Reduction of Labeled Data

Date
Abstract
Links
Bib
@article{BunteESANNSI2009,
author = {Kerstin Bunte and Barbara Hammer and Axel Wism\"uller and Michael Biehl},
title = {{Adaptive Local Dissimilarity Measures for Discriminative Dimension Reduction of Labeled Data}},
journal = {Neurocomputing},
month = {March},
volume = {73},
number = {7-9},
pages = {1074--1092},
year = {2010},
issn = {0925-2312},
numpages = {19},
editor = {Cecilio Angulo, John Lee, Frank-Michael Schleif},
acmid = {1751850},
publisher = {Elsevier Science Publishers B. V.},
address = {Amsterdam, The Netherlands, The Netherlands},
keywords = {Dimension reduction, Learning vector quantization, Visualization},
url = {http://dx.doi.org/10.1016/j.neucom.2009.11.017},
doi = {10.1016/j.neucom.2009.11.017},
abstract = {Due to the tremendous increase of electronic information with respect to the size of data sets as well as their dimension,  dimension reduction and visualization of high-dimensional data has become one of the key problems of data mining.  Since embedding in lower dimensions necessarily includes a loss of information, methods to explicitly control the information kept by a specific dimension reduction technique are highly desirable.  The incorporation of supervised class information constitutes an important specific case. The aim is to preserve and potentially enhance the discrimination of classes in lower dimensions.  In this contribution we use an extension of prototype-based local distance learning, which results in a nonlinear discriminative dissimilarity measure for a given labeled data manifold.  The learned local distance measure can be used as basis for other unsupervised dimension reduction techniques, which take into account neighborhood information.  We show the combination of different dimension reduction techniques with a discriminative similarity measure learned by an extension of learning vector quantization (LVQ)  and their behavior with different parameter settings. The methods are introduced and discussed in terms of artificial and real world data sets},
}