Dataset Viewer
The dataset could not be loaded because the splits use different data file formats, which is not supported. Read more about the splits configuration. Click for more details.
Couldn't infer the same data file format for all splits. Got {NamedSplit('train'): ('json', {}), NamedSplit('validation'): (None, {}), NamedSplit('test'): (None, {})}
Error code: FileFormatMismatchBetweenSplitsError
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
OpenWebText Expert Selections (GPT-OSS 20B)
This dataset contains top-k expert selections for each token from the first
200,000,000 tokens of vietgpt/openwebtext_en, using the router logits from
openai/gpt-oss-20b. Sequences are chunked to a maximum length of 32 tokens
within each document (no cross-document continuity); shorter tail chunks are
included without padding.
Files
openwebtext_200m_idx.npy: uint16 indices, shape(200_000_000, 24, 4)openwebtext_200m_val.npy: float16 router logits, shape(200_000_000, 24, 4)openwebtext_200m.meta.json: run metadata (model, dataset, seq_len, topk, etc.)openwebtext_200m.state.json: resume state (tokens_written, dataset offsets)collect_experts.py: collection script used to generate the files
Notes
- Top-k values are raw router logits (pre-softmax), not probabilities.
- Tokenization uses the model tokenizer (
openai/gpt-oss-20b) withadd_special_tokens=False.
Evaluation (text.txt, 2,584 tokens)
- MLP (L=5, 50k steps): top1=0.4822, top5=0.6649, top10=0.7078
- Transformer v2 (L=10, 25k steps): top1=0.7663, top5=0.8266, top10=0.8402
Eval command (no generation)
uv run python generate_and_eval.py \
--text-file text.txt \
--max-new-tokens 0 \
--seq-len 32 \
--layers 10 \
--mlp-layers 5 \
--tx-layers 10 \
--mlp-ckpt inverter_mlp_l5_50k.pt \
--tx-ckpt inverter_v2_l10_accum.pt \
--tx-d-model 768 \
--tx-n-layer 6 \
--tx-n-head 12 \
--tx-d-ff 2048 \
--layer-gating \
--attn-impl flash_attention_2 \
--out user_text_eval.json
Loading (memmap)
import numpy as np
idx = np.memmap(
"openwebtext_200m_idx.npy",
mode="r",
dtype=np.uint16,
shape=(200_000_000, 24, 4),
)
val = np.memmap(
"openwebtext_200m_val.npy",
mode="r",
dtype=np.float16,
shape=(200_000_000, 24, 4),
)
Collection command (for reference)
uv run python collect_experts.py \
--dataset vietgpt/openwebtext_en \
--model openai/gpt-oss-20b \
--topk 4 \
--seq-len 32 \
--batch-size 32 \
--max-tokens 200000000 \
--save-every 1000000 \
--attn-impl flash_attention_2 \
--output openwebtext_200m.npy \
--hard-exit
- Downloads last month
- 5