Co-EM Support Vector learning
Research output: Contributions to collected editions/works › Article in conference proceedings › Research › peer-review
Authors
Multi-view algorithms, such as co-training and co-EM, utilize unlabeled data when the available attributes can be split into independent and compatible subsets. Co-EM outperforms co-training for many problems, but it requires the underlying learner to estimate class probabilities, and to learn from probabilistically labeled data. Therefore, co-EM has so far only been studied with naive Bayesian learners. We cast linear classifiers into a probabilistic framework and develop a co-EM version of the Support Vector Machine. We conduct experiments on text classification problems and compare the family of semi-supervised support vector algorithms under different conditions, including violations of the assumptions underlying multiview learning. For some problems, such as course web page classification, we observe the most accurate results reported so far.
Original language | English |
---|---|
Title of host publication | Proceeding ICML '04 Proceedings of the twenty-first international conference on Machine learning |
Number of pages | 8 |
Place of Publication | New York |
Publisher | Association for Computing Machinery, Inc |
Publication date | 2004 |
Pages | 121-128 |
ISBN (print) | 1-58113-838-5 , 978-1-58113-838-2 |
DOIs | |
Publication status | Published - 2004 |
Event | 21st International Conference on Machine Learning - 2004 - Banff, Canada Duration: 31.12.2004 → … Conference number: 21 https://icml.cc/imls/icml.html |
- Informatics
- Business informatics