A topological reinforcement learning agent for navigation |
| |
Authors: | Arthur P S Braga Aluízio F R Araújo |
| |
Affiliation: | 1. Departamento de Engenharia Elétrica, Universidade de S?o Paulo, Av. Trabalhador S?ocarlense, 400, 13566-590, S?o Carlos, SP, Brazil
|
| |
Abstract: | This article proposes a reinforcement learning procedure for mobile robot navigation using a latent-like learning schema. Latent learning refers to learning that occurs in the absence of reinforcement signals and is not apparent until reinforcement is introduced. This concept considers that part of a task can be learned before the agent receives any indication of how to perform such a task. In the proposed topological reinforcement learning agent (TRLA), a topological map is used to perform the latent learning. The propagation of the reinforcement signal throughout the topological neighborhoods of the map permits the estimation of a value function which takes in average less trials and with less updatings per trial than six of the main temporal difference reinforcement learning algorithms: Q-learning, SARSA, Q(λ)-learning, SARSA(λ), Dyna-Q and fast Q(λ)-learning. The RL agents were tested in four different environments designed to consider a growing level of complexity in accomplishing navigation tasks. The tests suggested that the TRLA chooses shorter trajectories (in the number of steps) and/or requires less value function updatings in each trial than the other six reinforcement learning (RL) algorithms. |
| |
Keywords: | |
本文献已被 SpringerLink 等数据库收录! |
|