TU Darmstadt / ULB / TUbiblio

Learning to Approach a Moving Ball with a Simulated Two-Wheeled Robot

Flentge, Felix
Hrsg.: Bredenfeld, A. ; Jacoff, A. ; Noda, I. ; Takahashi, Y. (2006)
Learning to Approach a Moving Ball with a Simulated Two-Wheeled Robot.
Konferenzveröffentlichung, Bibliographie

Kurzbeschreibung (Abstract)

We show how a two-wheeled robot can learn to approach a moving ball using Reinforcement Learning. The robot is controlled by setting the velocities of its two wheels. It has to reach the ball under certain conditions to be able to kick it towards a given target. In order to kick, the ball has to be in front of the robot. The robot also has to reach the ball at a certain angle in relation to the target, because the ball is always kicked in the direction from the center of the robot to the ball. The robot learns which velocity differences should be applied to the wheels: one of the wheels is set to the maximum velocity, the other one according to this difference.We apply a REINFORCE algorithm [1] in combination with some kind of extended Growing Neural Gas (GNG) [2] to learn these continuous actions. The resulting algorithm, called ReinforceGNG, is tested in a simulated environment with and without noise.

Typ des Eintrags: Konferenzveröffentlichung
Erschienen: 2006
Herausgeber: Bredenfeld, A. ; Jacoff, A. ; Noda, I. ; Takahashi, Y.
Autor(en): Flentge, Felix
Art des Eintrags: Bibliographie
Titel: Learning to Approach a Moving Ball with a Simulated Two-Wheeled Robot
Sprache: Deutsch
Publikationsjahr: 2006
Verlag: Springer
Buchtitel: RoboCup 2005: Robot Soccer World Cup IX
Kurzbeschreibung (Abstract):

We show how a two-wheeled robot can learn to approach a moving ball using Reinforcement Learning. The robot is controlled by setting the velocities of its two wheels. It has to reach the ball under certain conditions to be able to kick it towards a given target. In order to kick, the ball has to be in front of the robot. The robot also has to reach the ball at a certain angle in relation to the target, because the ball is always kicked in the direction from the center of the robot to the ball. The robot learns which velocity differences should be applied to the wheels: one of the wheels is set to the maximum velocity, the other one according to this difference.We apply a REINFORCE algorithm [1] in combination with some kind of extended Growing Neural Gas (GNG) [2] to learn these continuous actions. The resulting algorithm, called ReinforceGNG, is tested in a simulated environment with and without noise.

Fachbereich(e)/-gebiet(e): 20 Fachbereich Informatik > Telekooperation
20 Fachbereich Informatik
Hinterlegungsdatum: 31 Dez 2016 12:59
Letzte Änderung: 15 Mai 2018 12:01
PPN:
Export:
Suche nach Titel in: TUfind oder in Google
Frage zum Eintrag Frage zum Eintrag

Optionen (nur für Redakteure)
Redaktionelle Details anzeigen Redaktionelle Details anzeigen