TY - JOUR
T1 - Predicting nearly as well as the best pruning of a decision tree through dynamic programming scheme
AU - Takimoto, Eiji
AU - Maruoka, Akira
AU - Vovk, Volodya
N1 - Funding Information:
The authors are grateful to Yoram Singer and an anonymous referee for their careful read ing of the paper and their helpful comments. This work was supported in part by Grant-in-Aid for Scienti/c Research on Priority Areas “Discovery Science” from the Ministry of Education, Science and Culture, Japan.
PY - 2001
Y1 - 2001
N2 - Helmbold and Schapire gave an on-line prediction algorithm that, when given an unpruned decision tree, produces predictions not much worse than the predictions made by the best pruning of the given decision tree. In this paper, we give two new on-line algorithms. The first algorithm is based on the observation that finding the best pruning can be efficiently solved by a dynamic programming in the "batch" setting where all the data to be predicted are given in advance. This algorithm works well for a wide class of loss functions, whereas the one given by Helmbold and Schapire is only described for the absolute loss function. Moreover, the algorithm given in this paper is so simple and general that it could be applied to many other on-line optimization problems solved by dynamic programming. We also explore the second algorithm that is competitive not only with the best pruning but also with the best prediction values which are associated with nodes in the decision tree. In this setting, a greatly simplified algorithm is given for the absolute loss function. It can be easily generalized to the case where, instead of using decision trees, data are classified in some arbitrarily fixed manner.
AB - Helmbold and Schapire gave an on-line prediction algorithm that, when given an unpruned decision tree, produces predictions not much worse than the predictions made by the best pruning of the given decision tree. In this paper, we give two new on-line algorithms. The first algorithm is based on the observation that finding the best pruning can be efficiently solved by a dynamic programming in the "batch" setting where all the data to be predicted are given in advance. This algorithm works well for a wide class of loss functions, whereas the one given by Helmbold and Schapire is only described for the absolute loss function. Moreover, the algorithm given in this paper is so simple and general that it could be applied to many other on-line optimization problems solved by dynamic programming. We also explore the second algorithm that is competitive not only with the best pruning but also with the best prediction values which are associated with nodes in the decision tree. In this setting, a greatly simplified algorithm is given for the absolute loss function. It can be easily generalized to the case where, instead of using decision trees, data are classified in some arbitrarily fixed manner.
UR - http://www.scopus.com/inward/record.url?scp=0034925309&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=0034925309&partnerID=8YFLogxK
U2 - 10.1016/S0304-3975(00)00138-9
DO - 10.1016/S0304-3975(00)00138-9
M3 - Conference article
AN - SCOPUS:0034925309
SN - 0304-3975
VL - 261
SP - 179
EP - 209
JO - Theoretical Computer Science
JF - Theoretical Computer Science
IS - 1
T2 - 8th International Workshop on Algorithmic Learning Theory
Y2 - 6 October 1997 through 8 October 1997
ER -