Counting 3,663 Big Data & Machine Learning Frameworks, Toolsets, and Examples...
Suggestion? Feedback? Tweet @stkim1

Author
Last Commit
Mar. 26, 2019
Created
Jan. 30, 2017

ChainerRL

Build Status Coverage Status Documentation Status PyPI

ChainerRL is a deep reinforcement learning library that implements various state-of-the-art deep reinforcement algorithms in Python using Chainer, a flexible deep learning framework.

Breakout Humanoid Grasping

Installation

ChainerRL is tested with Python 2.7+ and 3.5.1+. For other requirements, see requirements.txt.

ChainerRL can be installed via PyPI:

pip install chainerrl

It can also be installed from the source code:

python setup.py install

Refer to Installation for more information on installation.

Getting started

You can try ChainerRL Quickstart Guide first, or check the examples ready for Atari 2600 and Open AI Gym.

For more information, you can refer to ChainerRL's documentation.

Algorithms

Algorithm Discrete Action Continous Action Recurrent Model CPU Async Training
DQN (including DoubleDQN etc.) ✓ (NAF) x
Categorical DQN x x
IQN x x x
DDPG x x
A3C
ACER
NSQ (N-step Q-learning) ✓ (NAF)
PCL (Path Consistency Learning)
PPO x x
TRPO x x

Following algorithms have been implemented in ChainerRL:

  • A3C (Asynchronous Advantage Actor-Critic)
  • ACER (Actor-Critic with Experience Replay)
  • Asynchronous N-step Q-learning
  • Categorical DQN
  • IQN
  • DQN (including Double DQN, Persistent Advantage Learning (PAL), Double PAL, Dynamic Policy Programming (DPP))
  • DDPG (Deep Deterministic Policy Gradients) (including SVG(0))
  • PGT (Policy Gradient Theorem)
  • PCL (Path Consistency Learning)
  • PPO (Proximal Policy Optimization)
  • TRPO (Trust Region Policy Optimization)

Q-function based algorithms such as DQN can utilize a Normalized Advantage Function (NAF) to tackle continuous-action problems as well as DQN-like discrete output networks.

Paper Implementations

The following papers have been implemented in ChainerRL:

Visualization

ChainerRL has a set of accompanying visualization tools in order to aid developers' ability to understand and debug their RL agents. With this visualization tool, the behavior of ChainerRL agents can be easily inspected from a browser UI.

Environments

Environments that support the subset of OpenAI Gym's interface (reset and step methods) can be used.

Contributing

Any kind of contribution to ChainerRL would be highly appreciated! If you are interested in contributing to ChainerRL, please read CONTRIBUTING.md.

License

MIT License.

Latest Releases
v0.6.0
 Feb. 28 2019
v0.5.0
 Nov. 15 2018
v0.4.0
 Jul. 23 2018
v0.3.0
 Dec. 8 2017
v0.2.0
 Jun. 7 2017