- (Domingos, 1999) ⇒ Pedro Domingos. (1999). “MetaCost: A General Method for Making Classifiers Cost-sensitive.” In: Proceedings of the fifth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ISBN:1-58113-143-7 doi:10.1145/312129.312220
Subject Headings: MetaCost
Research in machine learning, statistics and related fields has produced a wide variety of algorithms for classification. However, most of these algorithms assume that all errors have the same cost, which is seldom the case in KDD problems. Individually making each classification learner cost-sensitive is laborious, and often non-trivial. In this paper we propose a principled method for making an arbitrary classifier cost-sensitive by wrapping a cost-minimizing procedure around it. This procedure, called MetaCost, treats the underlying classifier as a black box, requiring no knowledge of its functioning or change to it. Unlike stratification, MetaCost is applicable to any number of classes and to arbitrary cost matrices. Empirical trials on a large suite of benchmark databases show that MetaCost almost always produces large cost reductions compared to the cost-blind classifier used (C4.5RULES) and to two forms of stratification. Further tests identify the key components of MetaCost and those that can be varied without substantial loss. Experiments on a larger database indicate that MetaCost scales well.
|1999 MetaCostAGeneralMethodforMaking||Pedro Domingos||MetaCost: A General Method for Making Classifiers Cost-sensitive||10.1145/312129.312220||1999|