A Feature Selection Technique based on Distributional Differences

Sung-Dong Kim
Volume: 2, No: 1, Page: 23 ~ 27, Year: 2006

Keywords: Feature Selection, Distributional Differences
Full Text:

Abstract
This paper presents a feature selection technique based on distributional differences for efficient machine learning. Initial training data consists of data including many features and a target value. We classified them into positive and negative data based on the target value. We then divided the range of the feature values into 10 intervals and calculated the distribution of the intervals in each positive and negative data. Then, we selected the features and the intervals of the features for which the distributional differences are over a certain threshold. Using the selected intervals and features, we could obtain the reduced training data. In the experiments, we will show that the reduced training data can reduce the training time of the neural network by about 40%, and we can obtain more profit on simulated stock trading using the trained functions as well.

Article Statistics
Multiple requests among the same broswer session are counted as one view (or download).
If you mouse over a chart, a box will show the data point's value.


Cite this article
IEEE Style
Sung-Dong Kim, "A Feature Selection Technique based on Distributional Differences," Journal of Information Processing Systems, vol. 2, no. 1, pp. 23~27, 2006. DOI: .

ACM Style
Sung-Dong Kim, "A Feature Selection Technique based on Distributional Differences," Journal of Information Processing Systems, 2, 1, (2006), 23~27. DOI: .