Abstract
In the existing Reinforcement Learning, it is difficult and time consuming to find appropriate the meta-parameters such as learning rate, eligibility traces and temperature for exploration, in particular on a complicated and large-scale problem, the delayed reward often occurs and causes a difficulty in solving the problem. In this paper, we propose a novel method introducing a temperature distribution for reinforcement learning. In addition to the acquirement of policy based on profit sharing, the temperature is given to each state and is trained by hill-climbing method using likelihood function based on success and failure of the task. The proposed method can reduce the parameter setting according to the given problems. We showed the performance on the grid world problem and the control of Acrobot.
Original language | English |
---|---|
Pages (from-to) | 297-305 |
Number of pages | 9 |
Journal | Transactions of the Japanese Society for Artificial Intelligence |
Volume | 20 |
Issue number | 4 |
DOIs | |
Publication status | Published - 2005 |
Keywords
- Delayed reward
- Maximum likelihood estimation
- Meta-parameter control
- Profit sharing
- Reinforcement Learning
- Temperature distribution
ASJC Scopus subject areas
- Artificial Intelligence