Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

Awesome Lists | Featured Topics | Projects

https://github.com/facebookresearch/ReAgent

A platform for Reasoning systems (Reinforcement Learning, Contextual Bandits, etc.)
https://github.com/facebookresearch/ReAgent

Last synced: about 2 months ago
JSON representation

A platform for Reasoning systems (Reinforcement Learning, Contextual Bandits, etc.)

Awesome Lists containing this project

README

        

![Banner](logo/reagent_banner.png)

### ReAgent is officially archived and no longer maintained. For latest support on production-ready reinforcement learning open-source library, please refer to [Pearl](https://github.com/facebookresearch/Pearl/) - Production-ready Reinforcement Learning AI Agent Library, by the Applied Reinforcement Learning team @ Meta.

[![Support Ukraine](https://img.shields.io/badge/Support-Ukraine-FFD500?style=flat&labelColor=005BBB)](https://opensource.fb.com/support-ukraine)
[![License](https://img.shields.io/badge/license-BSD%203--Clause-brightgreen)](LICENSE)
[![CircleCI](https://circleci.com/gh/facebookresearch/ReAgent/tree/main.svg?style=shield)](https://circleci.com/gh/facebookresearch/ReAgent/tree/main)
[![codecov](https://codecov.io/gh/facebookresearch/ReAgent/branch/main/graph/badge.svg)](https://codecov.io/gh/facebookresearch/ReAgent)
---

### Overview
ReAgent is an open source end-to-end platform for applied reinforcement learning (RL) developed and used at Facebook. ReAgent is built in Python and uses PyTorch for modeling and training and TorchScript for model serving. The platform contains workflows to train popular deep RL algorithms and includes data preprocessing, feature transformation, distributed training, counterfactual policy evaluation, and optimized serving. For more detailed information about ReAgent see the release post [here](https://research.fb.com/publications/horizon-facebooks-open-source-applied-reinforcement-learning-platform/) and white paper [here](https://arxiv.org/abs/1811.00260).

The platform was once named "Horizon" but we have adopted the name "ReAgent" recently to emphasize its broader scope in decision making and reasoning.

### Algorithms Supported

Classic Off-Policy algorithms:
- Discrete-Action [DQN](https://storage.googleapis.com/deepmind-media/dqn/DQNNaturePaper.pdf)
- Parametric-Action DQN
- [Double DQN](https://arxiv.org/abs/1509.06461), [Dueling DQN](https://arxiv.org/abs/1511.06581), [Dueling Double DQN](https://arxiv.org/abs/1710.02298)
- Distributional RL: [C51](https://arxiv.org/abs/1707.06887) and [QR-DQN](https://arxiv.org/abs/1710.10044)
- [Twin Delayed DDPG](https://arxiv.org/abs/1802.09477) (TD3)
- [Soft Actor-Critic](https://arxiv.org/abs/1801.01290) (SAC)
- [Critic Regularized Regression](https://arxiv.org/abs/2006.15134) (CRR)
- [Proximal Policy Optimization Algorithms](https://arxiv.org/abs/1707.06347) (PPO)

RL for recommender systems:
- [Seq2Slate](https://arxiv.org/abs/1810.02019)
- [SlateQ](https://arxiv.org/abs/1905.12767)

Counterfactual Evaluation:
- [Doubly Robust](https://arxiv.org/abs/1612.01205) (for bandits)
- [Doubly Robust](https://arxiv.org/abs/1511.03722) (for sequential decisions)
- [MAGIC](https://arxiv.org/abs/1604.00923)

Multi-Arm and Contextual Bandits:
- [UCB1](https://www.cs.bham.ac.uk/internal/courses/robotics/lectures/ucb1.pdf)
- [MetricUCB](https://arxiv.org/abs/0809.4882)
- [Thompson Sampling](https://web.stanford.edu/~bvr/pubs/TS_Tutorial.pdf)
- [LinUCB](https://arxiv.org/abs/1003.0146)

Others:
- [Cross-Entropy Method](http://web.mit.edu/6.454/www/www_fall_2003/gew/CEtutorial.pdf)
- [Synthetic Return for Credit Assignment](https://arxiv.org/abs/2102.12425)

### Installation
ReAgent can be installed via. Docker or manually. Detailed instructions on how to install ReAgent can be found [here](docs/installation.rst).

### Tutorial
ReAgent is designed for large-scale, distributed recommendation/optimization tasks where we don’t have access to a simulator.
In this environment, it is typically better to train offline on batches of data, and release new policies slowly over time.
Because the policy updates slowly and in batches, we use off-policy algorithms. To test a new policy without deploying it,
we rely on counter-factual policy evaluation (CPE), a set of techniques for estimating a policy based on the actions of another policy.

We also have a set of tools to facilitate applying RL in real-world applications:
- Domain Analysis Tool, which analyzes state/action feature importance and identifies whether the problem is a suitable for applying batch RL
- Behavior Cloning, which clones from the logging policy to bootstrap the learning policy safely

Detailed instructions on how to use ReAgent can be found [here](docs/usage.rst).

### License
ReAgent is released under a BSD 3-Clause license. Find out more about it [here](LICENSE).

[Terms of Use](https://opensource.facebook.com/legal/terms) | [Privacy Policy](https://opensource.facebook.com/legal/privacy) | Copyright © 2022 Meta Platforms, Inc

### Citing
```
@article{gauci2018horizon,
title={Horizon: Facebook's Open Source Applied Reinforcement Learning Platform},
author={Gauci, Jason and Conti, Edoardo and Liang, Yitao and Virochsiri, Kittipat and Chen, Zhengxing and He, Yuchen and Kaden, Zachary and Narayanan, Vivek and Ye, Xiaohui},
journal={arXiv preprint arXiv:1811.00260},
year={2018}
}
```