Publication | Closed Access
Classifier cascades and trees for minimizing feature evaluation cost
74
Citations
40
References
2014
Year
Artificial IntelligenceEngineeringMachine LearningMachine Learning ToolFeature ExtractionClassifier Evaluation CostImage AnalysisData ScienceData MiningPattern RecognitionDecision Tree LearningSupervised LearningClassifier CascadesMachine Learning ModelPredictive AnalyticsFeature Extraction CostKnowledge DiscoveryComputer ScienceDeep LearningClassifier SystemCost-sensitive Machine Learning
Machine learning algorithms have successfully entered industry through many real-world applications (e.g., search engines and product recommendations). In these applications, the test-time CPU cost must be budgeted and accounted for. In this paper, we examine two main components of the test-time CPU cost, classifier evaluation cost and feature extraction cost, and show how to balance these costs with the classifier accuracy. Since the computation required for feature extraction dominates the test-time cost of a classifier in these settings, we develop two algorithms to efficiently balance the performance with the test-time cost. Our first contribution describes how to construct and optimize a tree of classifiers, through which test inputs traverse along individual paths. Each path extracts different features and is optimized for a specific sub-partition of the input space. Our second contribution is a natural reduction of the tree of classifiers into a cascade. The cascade is particularly useful for class-imbalanced data sets as the majority of instances can be early-exited out of the cascade when the algorithm is sufficiently confident in its prediction. Because both approaches only compute features for inputs that benefit from them the most, we find our trained classifiers lead to high accuracies at a small fraction of the computational cost.
| Year | Citations | |
|---|---|---|
Page 1
Page 1