Skip to content

Desarrollo de algoritmos de aprendizaje por refuerzo con estados continuos

License

Notifications You must be signed in to change notification settings

angel-ayala/continuous-rl

Repository files navigation

Reinforcement learning using continuous states and interactive feedback

Angel Ayala, Claudio Henríquez, Francisco Cruz

Universidad Central de Chile

Santiago, Chile

Research in intelligent systems field has led to different learningmethods for machines to acquire knowledge, among them, reinforcement learning (RL). Given the problem of the time required to learn how to develop a problem, using RL this work tackles the interactive reinforcement learning (IRL) approach as a way of solution for the training of agents. Furthermore, this work also addresses the problem of continuous representations along with the interactive approach. In this regards, we have performed experiments with simulated environments using different representations in the state vector in order to show the efficiency of this approach under a certain probability of interaction. The obtained results in the simulated environments show a faster learning convergence when using continuous states and interactive feedback in comparison to discrete and autonomous reinforcement learning respectively.

The authors gratefully acknowledge partial support by Universidad Central de Chile under the research project CIP2017030.

DOI: https://doi.org/10.1145/3309772.3309801

Paper: Go to file

Here you can find five different RL agents for two different environments implemented from Gym, an OpenAI toolkit.


Experimental Setup

Updates

  • Improvements of the ContinuousDQN agent 06/2019.
  • Centered Cart for the ContinuousDQN agent 08/2019.
  • Some codes reorganization!.

CartPole-v1 enviroment

For this environments has been developing three agents

  • The first one, cartpole-discreto discrete_qlearning agent, use the BOXES method for discretization of the state space, storing the Q values in memory.
  • The others two, continuous_qlearning agents, uses the DQN and DDQN methods for Q values approximation.

In the continuous observation space, the agent must learn an approximation function to achieve a better generalization. The agent ContinuousQlearning was designed to handle continuous observation space.


Arcade Learning Environments

For the ALE environments, the Space Invaders is used.

  • One has been implemented with DQN for the RAM variants of the Space Invaders.
  • Another has been implemented with DQN but with a CNN for the Q values approximation.

This environments is still in research how optimize with the state-of-art

If you need more information, email me angel4ayala at gmail.com

Results

Discrete Results Continuous Results

Results of training of 50 agents for the environment CartPole-v1 discretization state vector with BOXES,and feedback probability of 0.3.

Results of training of 50 agents for environment CartPole-v1 with continuous representation and feedback probability of 0.3.

About

Desarrollo de algoritmos de aprendizaje por refuerzo con estados continuos

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages