Abstract
High-quality, labeled data is essential for successfully applying machine learning methods to real-world text classification problems. However, in many cases, the amount of labeled data is very small compared to that of the unlabeled, and labeling additional samples could be expensive and time consuming. Co-training algorithms, which make use of unlabeled data to improve classification, have proven to be very effective in such cases. Generally, co-training algorithms work by using two classifiers, trained on two different views of the data, to label large amounts of unlabeled data. Doing so can help minimize the human effort required for labeling new data, as well as improve classification performance. In this article, we propose an ensemble-based co-training approach that uses an ensemble of classifiers from different training iterations to improve labeling accuracy. This approach, which we call vertical ensemble, incurs almost no additional computational cost. Experiments conducted on six textual datasets show a significant improvement of over 45% in AUC compared with the original co-training algorithm.
Original language | American English |
---|---|
Article number | 21 |
Journal | ACM Transactions on Intelligent Systems and Technology |
Volume | 9 |
Issue number | 2 |
DOIs | |
State | Published - 1 Oct 2017 |
Keywords
- Co-training
- Ensemble
- Text classification
All Science Journal Classification (ASJC) codes
- Theoretical Computer Science
- Artificial Intelligence