TY - JOUR
T1 - Boosting over non-deterministic ZDDs
AU - Fujita, Takahiro
AU - Hatano, Kohei
AU - Takimoto, Eiji
N1 - Funding Information:
We thank Yasuo Tabei and Daisuke Kimura of RIKEN AIP for insightful discussion and pointing out relevant related work. This work is supported in part by JSPS KAKENHI Grant Number JP16J04621 , JP16K00305 and JP15H02667 , respectively.
Funding Information:
We thank Yasuo Tabei and Daisuke Kimura of RIKEN AIP for insightful discussion and pointing out relevant related work. This work is supported in part by JSPS KAKENHI Grant Number JP16J04621, JP16K00305 and JP15H02667, respectively.
PY - 2020/2/2
Y1 - 2020/2/2
N2 - We propose a new approach to large-scale machine learning, learning over compressed data: First compress the training data somehow and then employ various machine learning algorithms on the compressed data, with the hope that the computation time is significantly reduced when the training data is well compressed. As a first step toward this approach, we consider a variant of the Zero-Suppressed Binary Decision Diagram (ZDD) as the data structure for representing the training data, which is a generalization of the ZDD by incorporating non-determinism. For the learning algorithm to be employed, we consider a boosting algorithm called AdaBoost⁎ and its precursor AdaBoost. In this paper, we give efficient implementations of the boosting algorithms whose running times (per iteration) are linear in the size of the given ZDD.
AB - We propose a new approach to large-scale machine learning, learning over compressed data: First compress the training data somehow and then employ various machine learning algorithms on the compressed data, with the hope that the computation time is significantly reduced when the training data is well compressed. As a first step toward this approach, we consider a variant of the Zero-Suppressed Binary Decision Diagram (ZDD) as the data structure for representing the training data, which is a generalization of the ZDD by incorporating non-determinism. For the learning algorithm to be employed, we consider a boosting algorithm called AdaBoost⁎ and its precursor AdaBoost. In this paper, we give efficient implementations of the boosting algorithms whose running times (per iteration) are linear in the size of the given ZDD.
UR - http://www.scopus.com/inward/record.url?scp=85058165937&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85058165937&partnerID=8YFLogxK
U2 - 10.1016/j.tcs.2018.11.027
DO - 10.1016/j.tcs.2018.11.027
M3 - Article
AN - SCOPUS:85058165937
SN - 0304-3975
VL - 806
SP - 81
EP - 89
JO - Theoretical Computer Science
JF - Theoretical Computer Science
ER -