TU Darmstadt / ULB / TUbiblio

Aktionenlernen mit Selbstorganisierenden Karten und Reinforcement Learning

Flentge, Felix (2005)
Aktionenlernen mit Selbstorganisierenden Karten und Reinforcement Learning.
Johannes Gutenberg-Universität Mainz
Ph.D. Thesis, Bibliographie

Abstract

This doctoral thesis deals with the development of a function approximator and its application to methods for learning discrete and continuous actions:

1. A general function approximator ? Locally Weighted Interpolating Growing Neural Gas (LWIGNG) ? is developed from Growing Neural Gas (GNG). The topological neighbourhood structure is used for calculating interpolations between neighbouring neurons and for applying a local weighting scheme. The capabilities of this method are shown in several experiments, with special considerations given to changing target functions and changing input distributions.

2. To learn discrete actions LWIGNG is combined with Q-Learning forming the Q-LWIGNG method. The underlying GNG-algorithm has to be changed to take care of the special order of the input data in action learning. Q-LWIGNG achieves very good results in experiments with the pole balancing and the mountain car problems, and good results with the acrobot problem.

3. To learn continuous actions a REINFORCE algorithm is combined with LWIGNG forming the ReinforceGNG method. An actor-critic architecture is used for learning from delayed rewards. LWIGNG approximates both the state-value function and the policy. The policy is given by the situation dependent parameters of a normal distribution. ReinforceGNG is applied successfully to learn continuous actions of a simulated 2-wheeled robot which has to intercept a rolling ball under certain conditions.

Item Type: Ph.D. Thesis
Erschienen: 2005
Creators: Flentge, Felix
Type of entry: Bibliographie
Title: Aktionenlernen mit Selbstorganisierenden Karten und Reinforcement Learning
Language: German
Date: 2005
Place of Publication: Mainz
Abstract:

This doctoral thesis deals with the development of a function approximator and its application to methods for learning discrete and continuous actions:

1. A general function approximator ? Locally Weighted Interpolating Growing Neural Gas (LWIGNG) ? is developed from Growing Neural Gas (GNG). The topological neighbourhood structure is used for calculating interpolations between neighbouring neurons and for applying a local weighting scheme. The capabilities of this method are shown in several experiments, with special considerations given to changing target functions and changing input distributions.

2. To learn discrete actions LWIGNG is combined with Q-Learning forming the Q-LWIGNG method. The underlying GNG-algorithm has to be changed to take care of the special order of the input data in action learning. Q-LWIGNG achieves very good results in experiments with the pole balancing and the mountain car problems, and good results with the acrobot problem.

3. To learn continuous actions a REINFORCE algorithm is combined with LWIGNG forming the ReinforceGNG method. An actor-critic architecture is used for learning from delayed rewards. LWIGNG approximates both the state-value function and the policy. The policy is given by the situation dependent parameters of a normal distribution. ReinforceGNG is applied successfully to learn continuous actions of a simulated 2-wheeled robot which has to intercept a rolling ball under certain conditions.

Divisions: 20 Department of Computer Science
20 Department of Computer Science > Telecooperation
Date Deposited: 31 Dec 2016 12:59
Last Modified: 22 Nov 2023 11:28
PPN:
Export:
Suche nach Titel in: TUfind oder in Google
Send an inquiry Send an inquiry

Options (only for editors)
Show editorial Details Show editorial Details