TY - JOUR
T1 - ForesTexter
T2 - An efficient random forest algorithm for imbalanced text categorization
AU - Wu, Qingyao
AU - Ye, Yunming
AU - Zhang, Haijun
AU - Ng, Michael K.
AU - Ho, Shen Shyang
N1 - Funding Information:
Y. Ye’s research supported in part by National Commonweal Technology R&D Program of AQSIQ China under Grant No. 201310087, Shenzhen Strategic Emerging Industries Program under Grants No. JCYJ20130329142551746. H. Zhang’s research supported in part by the National Nature Science Foundation of China under Gran No. 61300209, the Shenzhen Foundation Research Fund under Grant No. JCY20120613115205826 and the Shenzhen Technology Innovation Program under Grant No. CXZZ20130319100919673. S.S. Ho’s research supported in part by AcRF Grant RG-41/12 and Singapore NTU Start-Up Grant.
PY - 2014/9
Y1 - 2014/9
N2 - In this paper, we propose a new random forest (RF) based ensemble method, ForesTexter, to solve the imbalanced text categorization problems. RF has shown great success in many real-world applications. However, the problem of learning from text data with class imbalance is a relatively new challenge that needs to be addressed. A RF algorithm tends to use a simple random sampling of features in building their decision trees. As a result, it selects many subspaces that contain few, if any, informative features for the minority class. Furthermore, the Gini measure for data splitting is considered to be skew sensitive and bias towards the majority class. Due to the inherent complex characteristics of imbalanced text datasets, learning RF from such data requires new approaches to overcome challenges related to feature subspace selection and cut-point choice while performing node splitting. To this end, we propose a new tree induction method that selects splits, both feature subspace selection and splitting criterion, for RF on imbalanced text data. The key idea is to stratify features into two groups and to generate effective term weighting for the features. One group contains positive features for the minority class and the other one contains the negative features for the majority class. Then, for feature subspace selection, we effectively select features from each group based on the term weights. The advantage of our approach is that each subspace contains adequate informative features for both minority and majority classes. One difference between our proposed tree induction method and the classical RF method is that our method uses Support Vector Machines (SVM) classifier to split the training data into smaller and more balance subsets at each tree node, and then successively retrains the SVM classifiers on the data partitions to refine the model while moving down the tree. In this way, we force the classifiers to learn from refined feature subspaces and data subsets to fit the imbalanced data better. Hence, the tree model becomes more robust for text categorization task with imbalanced dataset. Experimental results on various benchmark imbalanced text datasets (Reuters-21578, Ohsumed, and imbalanced 20 newsgroup) consistently demonstrate the effectiveness of our proposed ForesTexter method. The performance of our proposed approach is competitive against the standard random forest and different variants of SVM algorithms.
AB - In this paper, we propose a new random forest (RF) based ensemble method, ForesTexter, to solve the imbalanced text categorization problems. RF has shown great success in many real-world applications. However, the problem of learning from text data with class imbalance is a relatively new challenge that needs to be addressed. A RF algorithm tends to use a simple random sampling of features in building their decision trees. As a result, it selects many subspaces that contain few, if any, informative features for the minority class. Furthermore, the Gini measure for data splitting is considered to be skew sensitive and bias towards the majority class. Due to the inherent complex characteristics of imbalanced text datasets, learning RF from such data requires new approaches to overcome challenges related to feature subspace selection and cut-point choice while performing node splitting. To this end, we propose a new tree induction method that selects splits, both feature subspace selection and splitting criterion, for RF on imbalanced text data. The key idea is to stratify features into two groups and to generate effective term weighting for the features. One group contains positive features for the minority class and the other one contains the negative features for the majority class. Then, for feature subspace selection, we effectively select features from each group based on the term weights. The advantage of our approach is that each subspace contains adequate informative features for both minority and majority classes. One difference between our proposed tree induction method and the classical RF method is that our method uses Support Vector Machines (SVM) classifier to split the training data into smaller and more balance subsets at each tree node, and then successively retrains the SVM classifiers on the data partitions to refine the model while moving down the tree. In this way, we force the classifiers to learn from refined feature subspaces and data subsets to fit the imbalanced data better. Hence, the tree model becomes more robust for text categorization task with imbalanced dataset. Experimental results on various benchmark imbalanced text datasets (Reuters-21578, Ohsumed, and imbalanced 20 newsgroup) consistently demonstrate the effectiveness of our proposed ForesTexter method. The performance of our proposed approach is competitive against the standard random forest and different variants of SVM algorithms.
KW - Imbalanced classification
KW - Random forests
KW - Stratified sampling
KW - SVM
KW - Text categorization
UR - http://www.scopus.com/inward/record.url?scp=84905560384&partnerID=8YFLogxK
U2 - 10.1016/j.knosys.2014.06.004
DO - 10.1016/j.knosys.2014.06.004
M3 - Journal article
AN - SCOPUS:84905560384
SN - 0950-7051
VL - 67
SP - 105
EP - 116
JO - Knowledge-Based Systems
JF - Knowledge-Based Systems
ER -