Vertical ensemble co-training for text classification

Gilad Katz, Cornelia Caragea, Asaf Shabtai

Research output: Contribution to journalArticlepeer-review

Abstract

High-quality, labeled data is essential for successfully applying machine learning methods to real-world text classification problems. However, in many cases, the amount of labeled data is very small compared to that of the unlabeled, and labeling additional samples could be expensive and time consuming. Co-training algorithms, which make use of unlabeled data to improve classification, have proven to be very effective in such cases. Generally, co-training algorithms work by using two classifiers, trained on two different views of the data, to label large amounts of unlabeled data. Doing so can help minimize the human effort required for labeling new data, as well as improve classification performance. In this article, we propose an ensemble-based co-training approach that uses an ensemble of classifiers from different training iterations to improve labeling accuracy. This approach, which we call vertical ensemble, incurs almost no additional computational cost. Experiments conducted on six textual datasets show a significant improvement of over 45% in AUC compared with the original co-training algorithm.

Original languageAmerican English
Article number21
JournalACM Transactions on Intelligent Systems and Technology
Volume9
Issue number2
DOIs
StatePublished - 1 Oct 2017

Keywords

  • Co-training
  • Ensemble
  • Text classification

All Science Journal Classification (ASJC) codes

  • Theoretical Computer Science
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Vertical ensemble co-training for text classification'. Together they form a unique fingerprint.

Cite this