TY - GEN
T1 - ACE-Cost
T2 - 7th International Conference on Machine Learning and Data Mining, MLDM 2011
AU - Li, Liyun
AU - Topkara, Umut
AU - Memon, Nasir
PY - 2011
Y1 - 2011
N2 - The standard prediction process of SVM requires acquisition of all the feature values for every instance. In practice, however, a cost is associated with the mere act of acquisition of a feature, e.g. CPU time needed to compute the feature out of raw data, the dollar amount spent for gleaning more information, or the patient wellness sacrificed by an invasive medical test, etc. In such applications, a budget constrains the classification process from using all of the features. We present, AceCost, a novel classification method that reduces the expected test cost of SVM without compromising from the classification accuracy. Our algorithm uses a cost efficient decision tree to partition the feature space for obtaining coarse decision boundaries, and local SVM classifiers at the leaves of the tree to refine them. The resulting classifiers are also effective in scenarios where several features share overlapping acquisition procedures, hence the cost of acquiring them as a group is less than the sum of the individual acquisition costs. Our experiments on the standard UCI datasets, a network flow detection application, as well as on synthetic datasets show that, the proposed approach achieves classification accuracy of SVM while reducing the test cost by 40%-80%.
AB - The standard prediction process of SVM requires acquisition of all the feature values for every instance. In practice, however, a cost is associated with the mere act of acquisition of a feature, e.g. CPU time needed to compute the feature out of raw data, the dollar amount spent for gleaning more information, or the patient wellness sacrificed by an invasive medical test, etc. In such applications, a budget constrains the classification process from using all of the features. We present, AceCost, a novel classification method that reduces the expected test cost of SVM without compromising from the classification accuracy. Our algorithm uses a cost efficient decision tree to partition the feature space for obtaining coarse decision boundaries, and local SVM classifiers at the leaves of the tree to refine them. The resulting classifiers are also effective in scenarios where several features share overlapping acquisition procedures, hence the cost of acquiring them as a group is less than the sum of the individual acquisition costs. Our experiments on the standard UCI datasets, a network flow detection application, as well as on synthetic datasets show that, the proposed approach achieves classification accuracy of SVM while reducing the test cost by 40%-80%.
KW - Cost Efficient Classification
KW - Decision Tree
KW - Postpruning
KW - SVM
UR - http://www.scopus.com/inward/record.url?scp=80052325722&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=80052325722&partnerID=8YFLogxK
U2 - 10.1007/978-3-642-23199-5_5
DO - 10.1007/978-3-642-23199-5_5
M3 - Conference contribution
AN - SCOPUS:80052325722
SN - 9783642231988
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 60
EP - 74
BT - Machine Learning and Data Mining in Pattern Recognition - 7th International Conference, MLDM 2011, Proceedings
Y2 - 30 August 2011 through 3 September 2011
ER -