- Modified this commit created by Nikhil Barhate to accomodate the Franka Emika Panda Robot.
- Installed Panda-Gym
- Adapted the classes to account for the difference in return after the step
Open PPO_colab.ipynb
in Google Colab
to see original PPO implementation for roboschool
- See this link to review the details (learning rates, episode logging, utils, etc) of this implementation of PPO.
- New modifications include to the reward figure, the rendering, etc
- No changes in PPO
- To train a new network : run
train.py
- To test a preTrained network : run
test.py
- To plot graphs using log files : run
plot_graph.py
- To save images for gif and make gif using a preTrained network : run
make_gif.py
- All parameters and hyperparamters to control training / testing / graphs / gifs are in their respective
.py
file PPO_colab.ipynb
combines all the files in a jupyter-notebook- All the hyperparameters used for training (preTrained) policies are listed in the
README.md
in PPO_preTrained directory
- if the environment runs on CPU, use CPU as device for faster training.
Please use this bibtex if you want to cite this repository in your publications :
@misc{ppo_panda,
author = {Lobbezoo, Andrew},
title = {PyTorch Implementation of Proximal Policy Optimization for the OpenAI Panda},
year = {2022},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/alobbezoo/PPO-Panda}},
}
PPO Continuous PandaReachDense-v2 | PPO Continuous PandaReachDense-v2 |
---|---|
![]() |
![]() |
Trained and Tested on:
Python 3
PyTorch
NumPy
gym
Training Environments
gym
Graphs and gifs
pandas
matplotlib
Pillow
pyvirtualdisplay
python-opengl