Technical Report
 

Learning decision trees for loss minimization in multi-class problems

Público Deposited

Contenido Descargable

Descargar PDF
https://ir.library.oregonstate.edu/concern/technical_reports/p2676w73b

Descriptions

Attribute NameValues
Creator
Abstract
  • Many machine learning applications require classifiers that minimize an asymmetric loss function rather than the raw misclassification rate. We study methods for modifying C4.5 to incorporate arbitrary loss matrices. One way to incorporate loss information into C4.5 is to manipulate the weights assigned to the examples from different classes. For 2-class problems, this works for any loss matrix, but for k > 2 classes, it is not sufficient. Nonetheless, we ask what is the set of class weights that best approximates an arbitrary k x k loss matrix, and we test and compare several methods: a wrapper method and some simple heuristics. The best method is a wrapper method that directly optimizes the loss using a hold- out data set. We define complexity measure for loss matrices and show that this measure can predict when more efficient methods will suffice and when the wrapper method must be applied.
  • Keywords: supervised learning, decision trees, misclassification costs, function minimization, loss
  • Keywords: supervised learning, decision trees, misclassification costs, function minimization, loss
Resource Type
Fecha Disponible
Fecha de Emisión
Series
Subject
Declaración de derechos
Publisher
Peer Reviewed
Language
Replaces

Relaciones

Parents:

This work has no parents.

Elementos