Sciweavers

54 search results - page 4 / 11
» An Empirical Evaluation of Bagging and Boosting
Sort
View
JMLR
2010
130views more  JMLR 2010»
13 years 2 months ago
MOA: Massive Online Analysis, a Framework for Stream Classification and Clustering
Massive Online Analysis (MOA) is a software environment for implementing algorithms and running experiments for online learning from evolving data streams. MOA is designed to deal...
Albert Bifet, Geoff Holmes, Bernhard Pfahringer, P...
ICML
1999
IEEE
14 years 8 months ago
Lazy Bayesian Rules: A Lazy Semi-Naive Bayesian Learning Technique Competitive to Boosting Decision Trees
Lbr is a lazy semi-naive Bayesian classi er learning technique, designed to alleviate the attribute interdependence problem of naive Bayesian classi cation. To classify a test exa...
Zijian Zheng, Geoffrey I. Webb, Kai Ming Ting
MCS
2004
Springer
14 years 24 days ago
A Comparison of Ensemble Creation Techniques
We experimentally evaluate bagging and six other randomization-based approaches to creating an ensemble of decision-tree classifiers. Bagging uses randomization to create multipl...
Robert E. Banfield, Lawrence O. Hall, Kevin W. Bow...
BMCBI
2004
176views more  BMCBI 2004»
13 years 7 months ago
Boosting accuracy of automated classification of fluorescence microscope images for location proteomics
Background: Detailed knowledge of the subcellular location of each expressed protein is critical to a full understanding of its function. Fluorescence microscopy, in combination w...
Kai Huang, Robert F. Murphy
ICML
2003
IEEE
14 years 8 months ago
Boosting Lazy Decision Trees
This paper explores the problem of how to construct lazy decision tree ensembles. We present and empirically evaluate a relevancebased boosting-style algorithm that builds a lazy ...
Xiaoli Zhang Fern, Carla E. Brodley