抄録
The acrobot is a two-link robot, actuated only at the joint between the two links. It is a difficult task in reinforcement learning (RL) to control the acrobot because it has nonlinear dynamics and continuous state and action spaces. In this article, we discuss applying the RL to the task of balancing control of the acrobot. Our RL method has an architecture similar to the actor-critic model. The actor is a controller that yields a control signal for the current state. The critic predicts the expected return in the future. The actor and the critic are approximated by normalized Gaussian networks, which are trained by an on-line EM algorithm. We also introduce a new method to promote the critic's learning. Our computer simulation shows that our method is able to acquire fairly good control through a small number of learning trials.
本文言語 | 英語 |
---|---|
ページ(範囲) | 12-20 |
ページ数 | 9 |
ジャーナル | Systems and Computers in Japan |
巻 | 32 |
号 | 5 |
DOI | |
出版ステータス | 出版済み - 05-2001 |
外部発表 | はい |
All Science Journal Classification (ASJC) codes
- 理論的コンピュータサイエンス
- 情報システム
- ハードウェアとアーキテクチャ
- 計算理論と計算数学