Transfer learning through policy abstraction using learning vector quantization

Ahmad Afif Mohd Faudzi, Hirotaka Takano, Junichi Murata

研究成果: ジャーナルへの寄稿学術誌査読

1 被引用数 (Scopus)

抄録

Reinforcement learning (RL) enables an agent to find a solution to a problem by interacting with the environment. However, the learning process always starts from scratch and possibly takes a long time. Here, knowledge transfer between tasks is considered. In this paper, we argue that an abstraction can improve the transfer learning. Modified learning vector quantization (LVQ) that can manipulate its network weights is proposed to perform an abstraction, an adaptation and a precaution. At first, the abstraction is performed by extracting an abstract policy out of a learned policy which is acquired through conventional RL method, Q-learning. The abstract policy then is used in a new task as prior information. Here, the adaptation or policy learning as well as new task's abstract policy generating are performed using only a single operation. Simulation results show that the representation of acquired abstract policy is interpretable, that the modified LVQ successfully performs policy learning as well as generates abstract policy and that the application of generalized common abstract policy produces better results by more effectively guiding the agent when learning a new task.

本文言語英語
ページ(範囲)163-168
ページ数6
ジャーナルJournal of Telecommunication, Electronic and Computer Engineering
10
1-3
出版ステータス出版済み - 2018

!!!All Science Journal Classification (ASJC) codes

  • ハードウェアとアーキテクチャ
  • コンピュータ ネットワークおよび通信
  • 電子工学および電気工学

フィンガープリント

「Transfer learning through policy abstraction using learning vector quantization」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル