Abstract
We show how to train SVMs with an optimal guarantee on the number of support vectors (up to constants), and with sample complexity and training runtime bounds matching the best known for kernel SVM optimization (i.e. without any additional asymptotic cost beyond standard SVM training). Our method is simple to implement and works well in practice.
Original language | English |
---|---|
Pages | 266-274 |
Number of pages | 9 |
State | Published - 2013 |
Event | 30th International Conference on Machine Learning, ICML 2013 - Atlanta, GA, United States Duration: 16 Jun 2013 → 21 Jun 2013 |
Conference
Conference | 30th International Conference on Machine Learning, ICML 2013 |
---|---|
Country/Territory | United States |
City | Atlanta, GA |
Period | 16/06/13 → 21/06/13 |
All Science Journal Classification (ASJC) codes
- Human-Computer Interaction
- Sociology and Political Science