TY - JOUR
T1 - Safe semi-supervised learning based on weighted likelihood
AU - Kawakita, Masanori
AU - Takeuchi, Jun'ichi
N1 - Funding Information:
This work was partially supported by grants from the Japan Society for the Promotion of Science ( KAKENHI 19300051 , 21700308 , 24500018 and 25870503 ). The authors are grateful to Dr. Takafumi Kanamori and Dr. Hironori Fujisawa for their helpful comments. Dr. Kanamori pointed out the relationship of our problem with the statistical paradox, whereas Dr. Fujisawa gave us useful comments on DRESS II. The authors also thank Dr. Natalie Sokolovska for valuable discussion.
PY - 2014/5
Y1 - 2014/5
N2 - We are interested in developing a safe semi-supervised learning that works in any situation. Semi-supervised learning postulates that n ' unlabeled data are available in addition to n labeled data. However, almost all of the previous semi-supervised methods require additional assumptions (not only unlabeled data) to make improvements on supervised learning. If such assumptions are not met, then the methods possibly perform worse than supervised learning. Sokolovska, Cappé, and Yvon (2008) proposed a semi-supervised method based on a weighted likelihood approach. They proved that this method asymptotically never performs worse than supervised learning (i.e.,it is safe) without any assumption. Their method is attractive because it is easy to implement and is potentially general. Moreover, it is deeply related to a certain statistical paradox. However, the method of Sokolovska etal. (2008) assumes a very limited situation, i.e.,classification, discrete covariates, n ' → ∞ and a maximum likelihood estimator. In this paper, we extend their method by modifying the weight. We prove that our proposal is safe in a significantly wide range of situations as long as n ≤ n '. Further, we give a geometrical interpretation of the proof of safety through the relationship with the above-mentioned statistical paradox. Finally, we show that the above proposal is asymptotically safe even when n ' < n by modifying the weight. Numerical experiments illustrate the performance of these methods.
AB - We are interested in developing a safe semi-supervised learning that works in any situation. Semi-supervised learning postulates that n ' unlabeled data are available in addition to n labeled data. However, almost all of the previous semi-supervised methods require additional assumptions (not only unlabeled data) to make improvements on supervised learning. If such assumptions are not met, then the methods possibly perform worse than supervised learning. Sokolovska, Cappé, and Yvon (2008) proposed a semi-supervised method based on a weighted likelihood approach. They proved that this method asymptotically never performs worse than supervised learning (i.e.,it is safe) without any assumption. Their method is attractive because it is easy to implement and is potentially general. Moreover, it is deeply related to a certain statistical paradox. However, the method of Sokolovska etal. (2008) assumes a very limited situation, i.e.,classification, discrete covariates, n ' → ∞ and a maximum likelihood estimator. In this paper, we extend their method by modifying the weight. We prove that our proposal is safe in a significantly wide range of situations as long as n ≤ n '. Further, we give a geometrical interpretation of the proof of safety through the relationship with the above-mentioned statistical paradox. Finally, we show that the above proposal is asymptotically safe even when n ' < n by modifying the weight. Numerical experiments illustrate the performance of these methods.
UR - http://www.scopus.com/inward/record.url?scp=84896083313&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84896083313&partnerID=8YFLogxK
U2 - 10.1016/j.neunet.2014.01.016
DO - 10.1016/j.neunet.2014.01.016
M3 - Article
C2 - 24632000
AN - SCOPUS:84896083313
SN - 0893-6080
VL - 53
SP - 146
EP - 164
JO - Neural Networks
JF - Neural Networks
ER -