Using POMDPs for learning cost sensitive decision trees

Shlomi Maliah, Guy Shani

Research output: Contribution to journalArticlepeer-review

Abstract

In classification, an algorithm learns to classify a given instance based on a set of observed attribute values. In many real world cases testing the value of an attribute incurs a cost. Furthermore, there can also be a cost associated with the misclassification of an instance. Cost sensitive classification attempts to minimize the expected cost of classification, by deciding after each observed attribute value, which attribute to measure next. In this paper we suggest Partially Observable Markov Decision Processes (POMDPs) as a modeling tool for cost sensitive classification. POMDPs are typically solved through a policy over belief states. We show how a relatively small set of potentially important belief states can be identified, and define an MDP over these belief states. To identify these potentially important belief states, we construct standard decision trees over all attribute subsets, and the leaves of these trees become the state space of our tree-based MDP. At each phase we decide on the next attribute to measure, balancing the cost of the measurement and the classification accuracy. We compare our approach to a set of previous approaches, showing our approach to work better for a range of misclassification costs.

Original languageAmerican English
Article number103400
JournalArtificial Intelligence
Volume292
DOIs
StatePublished - 1 Mar 2021

Keywords

  • Cost sensitive classification
  • Decision trees
  • MDP
  • POMDP

All Science Journal Classification (ASJC) codes

  • Language and Linguistics
  • Linguistics and Language
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Using POMDPs for learning cost sensitive decision trees'. Together they form a unique fingerprint.

Cite this