Distributed Ppo Save

This is an pytorch implementation of Distributed Proximal Policy Optimization(DPPO).

Project README

Distributed Proximal Policy Optimization (DPPO)

MIT License
This is an pytorch-version implementation of Emergence of Locomotion Behaviours in Rich Environments. This project is based on Alexis David Jacq's DPPO project. However, it has been rewritten and contains some modifications that appaer to improve learning in some environments. In this code, I revised the Running Mean Filter and this leads to better performance (for example in Walker2D). I also rewrote the code to support the Actor Network and Critic Network separately. This change then allows the creation of asymmetric for some tasks, where the information available at training time is not available at run time. Further, the actions in this project are sampled from a Beta Distribution, leads to better training speed and performance in a large number of tasks.

Requirements

  • python 3.5.2
  • openai gym
  • mujoco-python
  • pytorch-0.3.1 (will update to the 0.4.1 version in August!)
  • pyro

Instruction to run the code

Train your models

cd /root-of-this-code/
python train_network.py

You could also try other mujoco's environments. This code has already pre-trained one mujoco environment: Walker2d-v1. You could try it by yourself on your favourite task!

Test your models:

cd /root-of-this-code/
python demo.py

Results

Training Curve

img

Demo: Walker2d-v1

img

Acknowledgement

Reference

[1] Emergence of Locomotion Behaviours in Rich Environments

Open Source Agenda is not affiliated with "Distributed Ppo" Project. README Source: TianhongDai/distributed-ppo
Stars
60
Open Issues
1
Last Commit
5 years ago
License
MIT

Open Source Agenda Badge

Open Source Agenda Rating