Enhance model card: Add pipeline tag, library name, paper/project links, usage example, and update license
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,10 +1,45 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
datasets:
|
| 4 |
- jakegrigsby/metamon-synthetic
|
| 5 |
- jakegrigsby/metamon-parsed-replays
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
---
|
|
|
|
| 7 |
Checkpoints from Metamon (v1) training runs.
|
| 8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
|
| 10 |
-
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
datasets:
|
| 3 |
- jakegrigsby/metamon-synthetic
|
| 4 |
- jakegrigsby/metamon-parsed-replays
|
| 5 |
+
license: apache-2.0
|
| 6 |
+
pipeline_tag: reinforcement-learning
|
| 7 |
+
library_name: amago
|
| 8 |
+
tags:
|
| 9 |
+
- pokemon
|
| 10 |
+
- game-ai
|
| 11 |
+
- offline-rl
|
| 12 |
+
- transformers
|
| 13 |
---
|
| 14 |
+
|
| 15 |
Checkpoints from Metamon (v1) training runs.
|
| 16 |
|
| 17 |
+
**Metamon** enables reinforcement learning (RL) research on [Pokémon Showdown](https://pokemonshowdown.com/) by providing:
|
| 18 |
+
|
| 19 |
+
1) A standardized suite of teams and opponents for evaluation.
|
| 20 |
+
2) A large dataset of RL trajectories "reconstructed" from real human battles.
|
| 21 |
+
3) Starting points for training imitation learning (IL) and RL policies.
|
| 22 |
+
|
| 23 |
+
Metamon is the codebase behind ["Human-Level Competitive Pokémon via Scalable Offline RL and Transformers"](https://arxiv.org/abs/2504.04395) (RLC, 2025). Please check out our [project website](https://metamon.tech) for an overview of our results. This README documents the dataset, pretrained models, training, and evaluation details to help you get battling!
|
| 24 |
+
|
| 25 |
+
**Paper:** [Human-Level Competitive Pokémon via Scalable Offline Reinforcement Learning with Transformers](https://huggingface.co/papers/2504.04395)
|
| 26 |
+
**Project Website:** https://metamon.tech
|
| 27 |
+
**Code:** [GitHub Repository](https://github.com/UT-Austin-RPL/metamon/tree/main)
|
| 28 |
+
|
| 29 |
+
### Usage
|
| 30 |
+
|
| 31 |
+
Pretrained models can run without research GPUs, but you will need to install [`amago`](https://github.com/UT-Austin-RPL/amago), which is an RL codebase by the same authors. Follow installation instructions [here](https://ut-austin-rpl.github.io/amago/installation.html).
|
| 32 |
+
|
| 33 |
+
Load and run pretrained models with `metamon.rl.eval_pretrained`. For example, to run the default checkpoint of the `SyntheticRLV2` model for 100 battles against a set of heuristic baselines:
|
| 34 |
+
|
| 35 |
+
```bash
|
| 36 |
+
python -m metamon.rl.eval_pretrained --agent SyntheticRLV2 --gens 1 --formats ou --n_challenges 100 --eval_type heuristic
|
| 37 |
+
```
|
| 38 |
+
|
| 39 |
+
To battle against other models or humans online (via a local Showdown server):
|
| 40 |
+
|
| 41 |
+
```bash
|
| 42 |
+
python -m metamon.rl.eval_pretrained --agent SyntheticRLV2 --gens 1 --formats ou --n_challenges 50 --eval_type ladder --username <pick unique username> --team_set competitive
|
| 43 |
+
```
|
| 44 |
|
| 45 |
+
For more details on models and usage, please refer to the [project's GitHub repository](https://github.com/UT-Austin-RPL/metamon/tree/main).
|