garage

A framework for reproducible reinforcement learning research

Github星跟蹤圖

Docs
Build Status
License
codecov
PyPI version

garage

garage is a toolkit for developing and evaluating reinforcement learning algorithms, and an accompanying library of state-of-the-art implementations built using that toolkit.

The toolkit provides wide range of modular tools for implementing RL algorithms, including:

  • Composable neural network models
  • Replay buffers
  • High-performance samplers
  • An expressive experiment definition interface
  • Tools for reproducibility (e.g. set a global random seed which all components respect)
  • Logging to many outputs, including TensorBoard
  • Reliable experiment checkpointing and resuming
  • Environment interfaces for many popular benchmark suites
  • Supporting for running garage in diverse environments, including always up-to-date Docker containers

See the latest documentation for getting started instructions and detailed APIs.

Installation

pip install garage

Algorithms

The table below summarizes the algorithms available in garage., Algorithm, Framework(s), ----------------------, -------------------, CEM, numpy, CMA-ES, numpy, REINFORCE (a.k.a. VPG), PyTorch, TensorFlow, DDPG, PyTorch, TensorFlow, DQN, TensorFlow, DDQN, TensorFlow, ERWR, TensorFlow, NPO, TensorFlow, PPO, PyTorch, TensorFlow, REPS, TensorFlow, TD3, TensorFlow, TNPG, TensorFlow, TRPO, PyTorch, TensorFlow, MAML, PyTorch, ## Supported Tools and Frameworks
garage supports Python 3.5+.

We currently support PyTorch and TensorFlow for implementing the neural network portions of RL algorithms, and additions of new framework support are always welcome. PyTorch modules can be found in the package garage.torch and TensorFlow modules can be found in the package garage.tf. Algorithms which do not require neural networks are found in the package garage.np.

The package is available for download on PyPI, and we ensure that it installs successfully into environments defined using conda, Pipenv, and virtualenv.

All components use the popular gym.Env interface for RL environments.

Testing

The most important feature of garage is its comprehensive automated unit test and benchmarking suite, which helps ensure that the algorithms and modules in garage maintain state-of-the-art performance as the software changes.

Our testing strategy has three pillars:

  • Automation:
    We use continuous integration to test all modules and algorithms in garage before adding any change. The full installation and test suite is also run nightly, to detect regressions.
  • Acceptance Testing:
    Any commit which might change the performance of an algorithm is subjected to comprehensive benchmarks on the relevant algorithms before it is merged
  • Benchmarks and Monitoring:
    We benchmark the full suite of algorithms against their relevant benchmarks and widely-used implementations regularly, to detect regressions and improvements we may have missed.

Supported Releases, Release, Build Status, Last date of support, -------, ------------, --------------------, v2019.10, Build Status, June 30th, 2019, Garage releases a new stable version approximately every 4 months, in February, June, and October. Maintenance releases have a stable API and dependency tree, and receive bug fixes and critical improvements but not new features. We currently support each release for a window of 8 months.

Citing garage

If you use garage for academic research, please cite the repository using the following BibTeX entry. You should update the commit field with the commit or release tag your publication uses.

@misc{garage,
 author = {The garage contributors},
 title = {Garage: A toolkit for reproducible reinforcement learning research},
 year = {2019},
 publisher = {GitHub},
 journal = {GitHub repository},
 howpublished = {\url{https://github.com/rlworkgroup/garage}},
 commit = {be070842071f736eb24f28e4b902a9f144f5c97b}
}

Credits

The original code for garage was adopted from predecessor project called rllab. The garage project is grateful for the contributions of the original rllab authors, and hopes to continue advancing the state of reproducibility in RL research in the same spirit.

rllab was developed by Rocky Duan (UC Berkeley/OpenAI), Peter Chen (UC Berkeley), Rein Houthooft (UC Berkeley/OpenAI), John Schulman (UC Berkeley/OpenAI), and Pieter Abbeel (UC Berkeley/OpenAI).

主要指標

概覽
名稱與所有者rlworkgroup/garage
主編程語言Python
編程語言Python (語言數: 5)
平台
許可證MIT License
所有者活动
創建於2018-06-10 21:31:23
推送於2023-05-04 14:44:22
最后一次提交2023-01-03 22:06:27
發布數33
最新版本名稱v2021.03.0 (發布於 )
第一版名稱v2018.10.0 (發布於 )
用户参与
星數2k
關注者數55
派生數316
提交數1.2k
已啟用問題?
問題數1012
打開的問題數202
拉請求數1136
打開的拉請求數33
關閉的拉請求數144
项目设置
已啟用Wiki?
已存檔?
是復刻?
已鎖定?
是鏡像?
是私有?