Previous work on feature weighting for case-based learning algorithms has tended to use either global weights or weights that vary over extremely local regions of the case space. This paper examines the use of coarsely local weighting schemes, where feature weights are allowed to vary but are identical for groups or clusters of cases. We present a new technique, called class distribution weighting CDW, that allows weights to vary at the class level. We further extend CDW into a family of related techniques that exhibit varying degrees of locality, from global to local. The class distribution techniques are then applied to a set of eleven concept learning tasks. We nd that one or more of the CDW variants signi cantly improves classi cation accuracy for nine of the eleven tasks. In addition, we nd that the relative importance of classes, features, and feature values in a particular domain determines which variant is most successful.