Sciweavers

ICCBR
1997
Springer

Examining Locally Varying Weights for Nearest Neighbor Algorithms

14 years 4 months ago
Examining Locally Varying Weights for Nearest Neighbor Algorithms
Previous work on feature weighting for case-based learning algorithms has tended to use either global weights or weights that vary over extremely local regions of the case space. This paper examines the use of coarsely local weighting schemes, where feature weights are allowed to vary but are identical for groups or clusters of cases. We present a new technique, called class distribution weighting CDW, that allows weights to vary at the class level. We further extend CDW into a family of related techniques that exhibit varying degrees of locality, from global to local. The class distribution techniques are then applied to a set of eleven concept learning tasks. We nd that one or more of the CDW variants signi cantly improves classi cation accuracy for nine of the eleven tasks. In addition, we nd that the relative importance of classes, features, and feature values in a particular domain determines which variant is most successful.
Nicholas Howe, Claire Cardie
Added 08 Aug 2010
Updated 08 Aug 2010
Type Conference
Year 1997
Where ICCBR
Authors Nicholas Howe, Claire Cardie
Comments (0)