Sciweavers

JIPS
2006

A Feature Selection Technique based on Distributional Differences

14 years 12 days ago
A Feature Selection Technique based on Distributional Differences
: This paper presents a feature selection technique based on distributional differences for efficient machine learning. Initial training data consists of data including many features and a target value. We classified them into positive and negative data based on the target value. We then divided the range of the feature values into 10 intervals and calculated the distribution of the intervals in each positive and negative data. Then, we selected the features and the intervals of the features for which the distributional differences are over a certain threshold. Using the selected intervals and features, we could obtain the reduced training data. In the experiments, we will show that the reduced training data can reduce the training time of the neural network by about 40%, and we can obtain more profit on simulated stock trading using the trained functions as well.
Sung-Dong Kim
Added 13 Dec 2010
Updated 13 Dec 2010
Type Journal
Year 2006
Where JIPS
Authors Sung-Dong Kim
Comments (0)