Suche
Close this search box.
Suche

Recurrent Neural State Estimationin Domains with Long-Term Dependencies

S. Duell, L. Weichbrodt, A. Hans, S. Udluft

Veröffentlicht:

2012, European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning

This paper presents a state estimation approach for reinforcement learning (RL) of a partially observable Markov decision process. It is based on a special recurrent neural network architecture, the Markov decision process extraction network with shortcuts (MPEN-S). In contrast to previous work regarding this topic, we address the problem of long-term dependencies, which cause major problems in many real-world applications. The architecture is designed to model the reward-relevant dynamics of an
environment and is capable to condense large sets of continuous observables to a compact Markovian state representation. The resulting estimate can be used as input for RL methods that assume the underlying system to be a Markov decision process. Although the approach was developed with RL in mind, it is also useful for general prediction tasks.