Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/kachayev/gym-microrts-paper-sb3
RL agent to play μRTS with Stable-Baselines3 and PyTorch
https://github.com/kachayev/gym-microrts-paper-sb3
gym-environment ppo pytorch real-time-strategy reinforcement-learning reinforcement-learning-agent
Last synced: 2 months ago
JSON representation
RL agent to play μRTS with Stable-Baselines3 and PyTorch
- Host: GitHub
- URL: https://github.com/kachayev/gym-microrts-paper-sb3
- Owner: kachayev
- Created: 2021-12-20T00:55:10.000Z (about 3 years ago)
- Default Branch: main
- Last Pushed: 2022-01-23T18:33:26.000Z (almost 3 years ago)
- Last Synced: 2023-03-12T03:06:17.031Z (almost 2 years ago)
- Topics: gym-environment, ppo, pytorch, real-time-strategy, reinforcement-learning, reinforcement-learning-agent
- Language: Python
- Homepage:
- Size: 293 KB
- Stars: 25
- Watchers: 2
- Forks: 4
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Gym-μRTS with Stable-Baselines3/PyTorch
This repo contains an attempt to reproduce Gridnet PPO with invalid action masking algorithm to play μRTS using [Stable-Baselines3](https://github.com/DLR-RM/stable-baselines3) library. Apart from reproducibility, this might open access to a diverse set of well tested algorithms, and toolings for training, evaluations, and more.
Original paper: [Gym-μRTS: Toward Affordable Deep Reinforcement Learning Research in Real-time Strategy Games](https://arxiv.org/abs/2105.13807).
Original code: [gym-microrts-paper](https://github.com/vwxyzjn/gym-microrts-paper).
![demo.gif](https://github.com/vwxyzjn/gym-microrts/raw/master/static/fullgame.gif)
## Install
Prerequisites:
* Python 3.7.1+
* Java 8.0+
* FFmpeg (for video capturing)```
git clone https://github.com/kachayev/gym-microrts-paper-sb3
cd gym-microrts-paper-sb3
python -m venv venv
source venv/bin/activate
pip install -r requirements.txt
```Note that I use newer version of `gym-microrts` compared to the one that was originally used for the paper.
## Training
To traing an agent:
```
$ python ppo_gridnet_diverse_encode_decode_sb3.py
```If everything is setup correctly, you'll see typicall SB3 verbose logging:
```
Using cuda device
---------------------------------
| microrts/ | |
| avg_exec_time | 0.00409 |
| num_calls | 256 |
| total_exec_time | 1.05 |
| time/ | |
| fps | 560 |
| iterations | 1 |
| time_elapsed | 10 |
| total_timesteps | 6144 |
---------------------------------
-----------------------------------------
| microrts/ | |
| avg_exec_time | 0.00321 |
| num_calls | 512 |
| total_exec_time | 1.64 |
| time/ | |
| fps | 164 |
| iterations | 2 |
| time_elapsed | 74 |
| total_timesteps | 12288 |
| train/ | |
| approx_kl | 0.001475019 |
| clip_fraction | 0.0575 |
| clip_range | 0.1 |
| entropy_loss | -1.46 |
| explained_variance | 0.00712 |
| learning_rate | 0.00025 |
| loss | 0.0579 |
| n_updates | 4 |
| policy_gradient_loss | -0.0032 |
| value_loss | 0.261 |
-----------------------------------------
```By default, all settings are set as close as possible to the original implementation from the paper as possible. Thought the script supports flexible params:
```shell
$ python ppo_gridnet_diverse_encode_decode_sb3.py \
--total-timesteps 10_000 \
--bot-envs coacAI=8 randomBiasedAI=8 \
--num-selfplay-envs 12 \
--batch-size 2048 \
--n-epochs 10
```A trained agent is automatically saved to `agents/` folder (or any other folder provided as `--exp-folder` parameter). Now you can use `enjoy.py` to test it out in action:
```shell
$ python enjoy.py \
--agent-file agents/ppo_gridnet_diverse_encode_decode_sb3__1__1640241051.zip \
--max-steps 1_000
--bot-envs randomBiasedAI=1
```Training progress is automatically logged to TensorBoard. Watch the progress locally:
```shell
$ tensorboard --logdir runs/
$ open http://localhost:6006
```To profile code use `cProfile`:
```shell
$ python -m cProfile -s cumulative enjoy.py \
--agent-file agents/ppo_gridnet_diverse_encode_decode_sb3__1__1640241051.zip \
--max-steps 4_000
--bot-envs workerRushAI=1
```As soon as correctness of the implementation is verified, I will provide details on how to use RL Baselines3 Zoo for training and evaluations.
## Implementational Caveats
A few notes / pain points regarding the implementation of the alrogithms, and the process of integrating it with stable-baselines3:
* Gym does not ship a space for "array of multidiscrete" use case (let's be honest, it's not very common). But it gives an option for defining your space when necessary. A new space, when defined, is not easy to integrate into SB3. In a few different places SB3 raises `NotImplementedError` facing unknown space ([example 1](https://github.com/DLR-RM/stable-baselines3/blob/df6f9de8f46509dad47e6d2e5620aa993b0fc883/stable_baselines3/common/distributions.py#L644), [example 2](https://github.com/DLR-RM/stable-baselines3/blob/df6f9de8f46509dad47e6d2e5620aa993b0fc883/stable_baselines3/common/preprocessing.py#L183)).
* Seems like switching to fully rolled out `MutliDiscrete` space definition has a significant performance penalty. Still investigating if this can be improved.
* Invalid masking is implemented by passing masks into observations from the wrapper (the observation space is replaced with `gym.spaces.Dict` to hold both observations and masks). By doing it this way, masks are now available for policy, and fit rollout buffer layout. Masking is implemented by setting logits into `-inf` (or to a rather small number).Look for `xxx(hack)` comments in the code for more details.
## More Experimentation
Additional experimentation with implementation details (those that are not present in the original paper) are now moved to separate scripts (to avoid confusion).
### Linear Critic
The idea is to have the critic (value approximation) to be done as an affine transformation rather than a 2-layers NN. In addition to the change, CNN output is now L2-normalized.
```shell
$ python ppo_gridnet_linear_critic.py \
--total-timesteps 10_000_000 \
--bot-envs coacAI=24 randomBiasedAI=24 \
--num-selfplay-envs 0 \
--batch-size 2048 \
--n-epochs 10
```### Linear Actor
After a quick analysis of embeddings space produced by encoder, some observations:
* enocder embeddings carry weak signal for reconstructing features of the environemnt (using linear probes)
* embeddings alongside a single trajectory do not exibit smoothnessHypothetically this means the encoder is "collapsed" with the actor network (decisions are made mostly on the encoder side). Practically this means weaker generalization. To test out the hypothesis, `ppo_gridnet_linear_actor` implements policy network as a simple linear controller applied to all cells on the map (leveraging the fact that encoder produces 256-dimensional vector).
```shell
$ python ppo_gridnet_linear_actor.py \
--total-timesteps 10_000_000 \
--bot-envs lightRushAI=12 workerRushAI=12 \
--num-selfplay-envs 0 \
--batch-size 2048 \
--n-epochs 10
```