nielsr HF Staff commited on
Commit
8cb3d7a
·
verified ·
1 Parent(s): b28a8b1

Enhance model card: Add pipeline tag, library name, paper/project links, usage example, and update license

Browse files

This PR enhances the model card by:

- Adding the `pipeline_tag: reinforcement-learning` to improve discoverability.
- Specifying `library_name: amago` as the primary library used by the model.
- Updating the license to `apache-2.0` as indicated by the project's GitHub.
- Including additional relevant tags like `pokemon`, `game-ai`, `offline-rl`, and `transformers`.
- Linking to the paper: [Human-Level Competitive Pokémon via Scalable Offline Reinforcement Learning with Transformers](https://huggingface.co/papers/2504.04395).
- Adding the project website: https://metamon.tech.
- Providing a concise overview and a practical Python usage example for quick model inference.

These changes will make the model more informative and easier to find and use for researchers and the community.

Files changed (1) hide show
  1. README.md +37 -2
README.md CHANGED
@@ -1,10 +1,45 @@
1
  ---
2
- license: mit
3
  datasets:
4
  - jakegrigsby/metamon-synthetic
5
  - jakegrigsby/metamon-parsed-replays
 
 
 
 
 
 
 
 
6
  ---
 
7
  Checkpoints from Metamon (v1) training runs.
8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
 
10
- Check out the project on [GitHub](https://github.com/UT-Austin-RPL/metamon/tree/main) for more information.
 
1
  ---
 
2
  datasets:
3
  - jakegrigsby/metamon-synthetic
4
  - jakegrigsby/metamon-parsed-replays
5
+ license: apache-2.0
6
+ pipeline_tag: reinforcement-learning
7
+ library_name: amago
8
+ tags:
9
+ - pokemon
10
+ - game-ai
11
+ - offline-rl
12
+ - transformers
13
  ---
14
+
15
  Checkpoints from Metamon (v1) training runs.
16
 
17
+ **Metamon** enables reinforcement learning (RL) research on [Pokémon Showdown](https://pokemonshowdown.com/) by providing:
18
+
19
+ 1) A standardized suite of teams and opponents for evaluation.
20
+ 2) A large dataset of RL trajectories "reconstructed" from real human battles.
21
+ 3) Starting points for training imitation learning (IL) and RL policies.
22
+
23
+ Metamon is the codebase behind ["Human-Level Competitive Pokémon via Scalable Offline RL and Transformers"](https://arxiv.org/abs/2504.04395) (RLC, 2025). Please check out our [project website](https://metamon.tech) for an overview of our results. This README documents the dataset, pretrained models, training, and evaluation details to help you get battling!
24
+
25
+ **Paper:** [Human-Level Competitive Pokémon via Scalable Offline Reinforcement Learning with Transformers](https://huggingface.co/papers/2504.04395)
26
+ **Project Website:** https://metamon.tech
27
+ **Code:** [GitHub Repository](https://github.com/UT-Austin-RPL/metamon/tree/main)
28
+
29
+ ### Usage
30
+
31
+ Pretrained models can run without research GPUs, but you will need to install [`amago`](https://github.com/UT-Austin-RPL/amago), which is an RL codebase by the same authors. Follow installation instructions [here](https://ut-austin-rpl.github.io/amago/installation.html).
32
+
33
+ Load and run pretrained models with `metamon.rl.eval_pretrained`. For example, to run the default checkpoint of the `SyntheticRLV2` model for 100 battles against a set of heuristic baselines:
34
+
35
+ ```bash
36
+ python -m metamon.rl.eval_pretrained --agent SyntheticRLV2 --gens 1 --formats ou --n_challenges 100 --eval_type heuristic
37
+ ```
38
+
39
+ To battle against other models or humans online (via a local Showdown server):
40
+
41
+ ```bash
42
+ python -m metamon.rl.eval_pretrained --agent SyntheticRLV2 --gens 1 --formats ou --n_challenges 50 --eval_type ladder --username <pick unique username> --team_set competitive
43
+ ```
44
 
45
+ For more details on models and usage, please refer to the [project's GitHub repository](https://github.com/UT-Austin-RPL/metamon/tree/main).