File size: 7,081 Bytes
e39576a 8459be0 e39576a a9f3e0e e39576a a9f3e0e e39576a a9f3e0e e39576a e4efec3 e39576a e4efec3 e39576a e4efec3 e39576a a9f3e0e e39576a a9f3e0e e39576a e4efec3 e39576a | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 | ---
license: cc-by-4.0
task_categories:
- text-generation
- reinforcement-learning
language:
- en
tags:
- chess
- uci
- transformer
- games
- elite
- lichess
- tokenized
size_categories:
- 1M<n<10M
---
# chess-elite-uci
A transformer-ready dataset of ~7.8 million elite chess games, pre-tokenized in UCI notation with a deterministic 1977-token vocabulary. Built for training chess language models directly with no preprocessing required.
## Dataset Summary
| Field | Value |
|---|---|
| Total games | 7,805,503 |
| Average sequence length | 94.24 tokens |
| Max sequence length | 255 tokens |
| Vocabulary size | 1,977 tokens |
| Mean combined Elo | 5,211 (~2,606 per player) |
## Sources
**Lichess Elite Database** (June 2020 – November 2025)
Games where both players are rated 2500+ vs 2300+ (2022 onwards: 2500+ vs 2300+; prior: 2400+ vs 2200+). Source: [database.nikonoel.fr](https://database.nikonoel.fr). Licensed CC0.
## Vocabulary
The vocabulary contains **1,977 tokens** and is fully deterministic and enumerated from chess geometry, not derived from data. It will never produce OOV tokens for any legal chess game.
| ID | Token | Description
|---|---|---|
| 0 | `<PAD>` | Padding
| 1 | `<W>` | POV token: white wins / white side for draws |
| 2 | `<B>` | POV token: black wins / black side for draws |
| 3 | `<CHECKMATE>` | Terminal: game ended in checkmate |
| 4 | `<RESIGN>` | Terminal: losing side resigned (≥ 40 ply) |
| 5 | `<STALEMATE>` | Terminal: draw by stalemate |
| 6 | `<REPETITION>` | Terminal: draw by threefold repetition |
| 7 | `<FIFTY_MOVE>` | Terminal: draw by 50-move rule |
| 8 | `<INSUFF_MATERIAL>` | Terminal: draw by insufficient material |
| 9+ | a1a2 … h7h8q | 1968 UCI move strings, sorted lexicographically |
The full vocabulary is provided in `vocab.json` as `{ token_str: int_id }`.
## Sequence Format
Every game is encoded as a flat list of integer token IDs:
```
[ <POV> | m1 | m2 | m3 | ... | mN | <TERMINAL> ]
```
- **POV token** (position 0): `<W>` if white wins, `<B>` if black wins. For draws, assigned randomly 50/50 between `<W>` and `<B>`.
- **Move tokens** (positions 1 to N): UCI half-moves alternating white/black, e.g. `e2e4`, `e7e5`, `g1f3`, `e1g1` (castling), `e7e8q` (promotion).
- **Terminal token** (position N+1): encodes why the game ended.
Maximum sequence length is **255 tokens** (1 POV + 253 moves + 1 terminal). Sequences are variable length, pad to 255 with `<PAD>` (ID 0) in your DataLoader.
## NTP Loss Mask
The `ntp_mask` column contains a binary list of the same length as `token_ids`. It indicates which positions should have next-token-prediction (NTP) loss applied during training:
```
Position NTP loss
─────────────────────────────
POV token 1 (always)
Winning side move 1
Losing side move 0 (context only)
Terminal token 1 (always)
Draw game moves 1 (both sides, since neither lost)
```
This implements win-conditioned training: the model learns to predict the winning side's moves given the POV token, while still attending to the losing side's moves as context.
Usage in PyTorch:
```python
loss = cross_entropy(logits, labels, reduction="none")
loss = (loss * ntp_mask).sum() / ntp_mask.sum()
```
## Filtering
Games were filtered as follows before inclusion:
**Decisive games (1-0 / 0-1):**
- **Checkmates**: verified by `board.is_checkmate()` on the final position. No length minimum.
- **Resignations**: not checkmate, minimum 40 halfmoves (20 moves each side).
**Draws (1/2-1/2):**
- Only **forced draws** are included: stalemate, insufficient material, 50-move rule, threefold repetition.
- Draw-by-agreement is excluded (`board.is_game_over(claim_draw=True)` must return True).
**All games:**
- Maximum 253 halfmoves (fits within 255-token sequence budget).
- Both player Elo values must be present and non-zero.
- All moves must be legally parseable by python-chess.
**Game type breakdown:**
| Type | Count | % |
|---|---|---|
| White checkmate | 1,702,751 | 21.8% |
| White resignation | 2,000,000 | 25.6% |
| Black checkmate | 1,702,752 | 21.8% |
| Black resignation | 2,000,000 | 25.6% |
| Forced draw | 400,000 | 5.1% |
## Schema
```python
{
"white_elo": int32, # white player Elo
"black_elo": int32, # black player Elo
"combined_elo": int32, # white_elo + black_elo
"result": string, # "1-0", "0-1", or "1/2-1/2"
"game_type": string, # "checkmate", "resignation", or "forced_draw"
"pov": string, # "<W>" or "<B>"
"terminal": string, # "<CHECKMATE>", "<RESIGN>", "<STALEMATE>", ...
"source": string, # "lichess"
"moves_uci": string, # space-separated UCI moves, human-readable
"token_ids": list[int32], # encoded sequence, use this for training
"ntp_mask": list[int32], # 1 = apply NTP loss, 0 = skip
"seq_len": int32, # len(token_ids), always in [3, 255]
}
```
## Usage
```python
from datasets import load_dataset
import json
# Load dataset
ds = load_dataset("MostLime/chess-elite-uci", split="train")
# Load vocabulary
with open("vocab.json") as f:
vocab = json.load(f)
id_to_token = {v: k for k, v in vocab.items()}
# Decode a game
row = ds[0]
tokens = [id_to_token[i] for i in row["token_ids"]]
print(" ".join(tokens))
# → <W> e2e4 e7e5 g1f3 b8c6 f1b5 ... <RESIGN>
# PyTorch DataLoader
import torch
from torch.utils.data import DataLoader
def collate(batch):
max_len = 255
token_ids = torch.zeros(len(batch), max_len, dtype=torch.long)
ntp_mask = torch.zeros(len(batch), max_len, dtype=torch.float)
for i, row in enumerate(batch):
n = row["seq_len"]
token_ids[i, :n] = torch.tensor(row["token_ids"], dtype=torch.long)
ntp_mask[i, :n] = torch.tensor(row["ntp_mask"], dtype=torch.float)
return {"token_ids": token_ids, "ntp_mask": ntp_mask}
loader = DataLoader(ds, batch_size=32, collate_fn=collate)
```
## Inference
At inference time, prepend the POV token for the side the model plays as, then feed opponent moves as context and sample responses:
```python
# Model plays as white
sequence = [vocab["<W>"]]
# Opponent plays e7e5 — append as context
sequence.append(vocab["e7e5"])
# Sample model's next move from legal UCI moves for the current position
```
Terminal tokens are never generated during normal play. The game ends when the opponent resigns or a draw is claimed externally.
## Citation
```bibtex
@dataset{mostlime2026chessEliteUCI,
author = {MostLime},
title = {chess-elite-uci: A Transformer-Ready Dataset of Elite Chess Games},
year = {2026},
publisher = {HuggingFace},
url = {https://huggingface.co/datasets/MostLime/chess-elite-uci}
}
```
## Acknowledgements
- [Lichess Elite Database](https://database.nikonoel.fr) by nikonoel — CC0
- [python-chess](https://python-chess.readthedocs.io) for move parsing and board state verification
- [Modal](https://modal.com) for distributed compute |