Skip to content

NYCU-MLLab/SRMRL-Skill-Regularized-Meta-Reinforcement-Learning

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

SRMRL: Skill Regularized Meta Reinforcement Learning

This is the implementation of SRMRL and this repository is based on garage.

Installation

To install locally, you will need to first install MuJoCo. Set LD_LIBRARY_PATH to point to both the MuJoCo binaries (/$HOME/.mujoco/mujoco200/bin) as well as the gpu drivers (something like /usr/lib/nvidia-390, you can find your version by running nvidia-smi).

Clone this repo and construct a virtual environment via pipenv install -r requirements.txt. Then activate the virtual environment with pipenv shell. The implementation of SRMRL is placed in garage/torch/algos/srmrl.py. Add the garage package into your python path with export PYTHONPATH=.:$PYTHONPATH in SRMRL directory.

Experiments

To run the experiments, you can use the scripts in script folder with

./script/[filename]

The script assumes you are in the SRMRL directory. Or you can run other tasks using

python example/[filename].py --env_name=[env_name]

The output files including log file and model parameters will be placed in ./data/local/experiment/[EXP_NAME]. The output log file can be visualized with tensorboard. The file progress.csv contains statistics logged over the course of training. We recommend viskit for visualizing learning curves: https://github.com/vitchyr/viskit.

After training, use sim_policy.py to visualize the learned policy:

python sim_policy.py --model_dir=[output_dir]

This script generate images and gif file of trajectories.

Visulization of learned polices

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%