TY - GEN
T1 - Active classification with comparison queries
AU - Kane, Daniel M.
AU - Lovett, Shachar
AU - Moran, Shay
AU - Zhang, Jiapeng
N1 - Publisher Copyright: © 2017 IEEE.
PY - 2017/11/10
Y1 - 2017/11/10
N2 - We study an extension of active learning in which the learning algorithm may ask the annotator to compare the distances of two examples from the boundary of their label-class. For example, in a recommendation system application (say for restaurants), the annotator may be asked whether she liked or disliked a specific restaurant (a label query); or which one of two restaurants did she like more (a comparison query).We focus on the class of half spaces, and show that under natural assumptions, such as large margin or bounded bit-description of the input examples, it is possible to reveal all the labels of a sample of size n using approximately O(log n) queries. This implies an exponential improvement over classical active learning, where only label queries are allowed. We complement these results by showing that if any of these assumptions is removed then, in the worst case, (n) queries are required.Our results follow from a new general framework of active learning with additional queries. We identify a combinatorial dimension, called the inference dimension, that captures the query complexity when each additional query is determined by O(1) examples (such as comparison queries, each of which is determined by the two compared examples). Our results for half spaces follow by bounding the inference dimension in the cases discussed above.
AB - We study an extension of active learning in which the learning algorithm may ask the annotator to compare the distances of two examples from the boundary of their label-class. For example, in a recommendation system application (say for restaurants), the annotator may be asked whether she liked or disliked a specific restaurant (a label query); or which one of two restaurants did she like more (a comparison query).We focus on the class of half spaces, and show that under natural assumptions, such as large margin or bounded bit-description of the input examples, it is possible to reveal all the labels of a sample of size n using approximately O(log n) queries. This implies an exponential improvement over classical active learning, where only label queries are allowed. We complement these results by showing that if any of these assumptions is removed then, in the worst case, (n) queries are required.Our results follow from a new general framework of active learning with additional queries. We identify a combinatorial dimension, called the inference dimension, that captures the query complexity when each additional query is determined by O(1) examples (such as comparison queries, each of which is determined by the two compared examples). Our results for half spaces follow by bounding the inference dimension in the cases discussed above.
UR - http://www.scopus.com/inward/record.url?scp=85041113234&partnerID=8YFLogxK
U2 - https://doi.org/10.1109/FOCS.2017.40
DO - https://doi.org/10.1109/FOCS.2017.40
M3 - منشور من مؤتمر
T3 - Annual Symposium on Foundations of Computer Science - Proceedings
SP - 355
EP - 366
BT - Proceedings - 58th Annual IEEE Symposium on Foundations of Computer Science, FOCS 2017
T2 - 58th Annual IEEE Symposium on Foundations of Computer Science, FOCS 2017
Y2 - 15 October 2017 through 17 October 2017
ER -