Flentge, Felix
Hrsg.: Bredenfeld, A. ; Jacoff, A. ; Noda, I. ; Takahashi, Y. (2006)
Learning to Approach a Moving Ball with a Simulated Two-Wheeled Robot.
Konferenzveröffentlichung, Bibliographie
Kurzbeschreibung (Abstract)
We show how a two-wheeled robot can learn to approach a moving ball using Reinforcement Learning. The robot is controlled by setting the velocities of its two wheels. It has to reach the ball under certain conditions to be able to kick it towards a given target. In order to kick, the ball has to be in front of the robot. The robot also has to reach the ball at a certain angle in relation to the target, because the ball is always kicked in the direction from the center of the robot to the ball. The robot learns which velocity differences should be applied to the wheels: one of the wheels is set to the maximum velocity, the other one according to this difference.We apply a REINFORCE algorithm [1] in combination with some kind of extended Growing Neural Gas (GNG) [2] to learn these continuous actions. The resulting algorithm, called ReinforceGNG, is tested in a simulated environment with and without noise.
Typ des Eintrags: | Konferenzveröffentlichung |
---|---|
Erschienen: | 2006 |
Herausgeber: | Bredenfeld, A. ; Jacoff, A. ; Noda, I. ; Takahashi, Y. |
Autor(en): | Flentge, Felix |
Art des Eintrags: | Bibliographie |
Titel: | Learning to Approach a Moving Ball with a Simulated Two-Wheeled Robot |
Sprache: | Deutsch |
Publikationsjahr: | 2006 |
Verlag: | Springer |
Buchtitel: | RoboCup 2005: Robot Soccer World Cup IX |
Kurzbeschreibung (Abstract): | We show how a two-wheeled robot can learn to approach a moving ball using Reinforcement Learning. The robot is controlled by setting the velocities of its two wheels. It has to reach the ball under certain conditions to be able to kick it towards a given target. In order to kick, the ball has to be in front of the robot. The robot also has to reach the ball at a certain angle in relation to the target, because the ball is always kicked in the direction from the center of the robot to the ball. The robot learns which velocity differences should be applied to the wheels: one of the wheels is set to the maximum velocity, the other one according to this difference.We apply a REINFORCE algorithm [1] in combination with some kind of extended Growing Neural Gas (GNG) [2] to learn these continuous actions. The resulting algorithm, called ReinforceGNG, is tested in a simulated environment with and without noise. |
Fachbereich(e)/-gebiet(e): | 20 Fachbereich Informatik > Telekooperation 20 Fachbereich Informatik |
Hinterlegungsdatum: | 31 Dez 2016 12:59 |
Letzte Änderung: | 15 Mai 2018 12:01 |
PPN: | |
Export: | |
Suche nach Titel in: | TUfind oder in Google |
Frage zum Eintrag |
Optionen (nur für Redakteure)
Redaktionelle Details anzeigen |