Datasets:
question
stringlengths 15
76
| choices
listlengths 4
4
| answer
int64 0
3
| answer_label
stringclasses 4
values | split
stringclasses 2
values | subcategories
stringclasses 2
values | lang
stringclasses 1
value | second_lang
stringclasses 1
value | notes
stringlengths 0
19
| id
stringlengths 7
7
| set_id
stringlengths 5
5
| variation_id
stringclasses 1
value | question_general_category
stringclasses 9
values | vanilla_cos_sim_to_canonical
dict | trimmed_cos_sim_to_canonical
dict | token_counts
dict |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
The markup language primarily used to create a web page is
|
[
"HTML",
"XHTML",
"JavaScript",
"CSS"
] | 0
|
A
|
dev
|
Canonical
|
eng_Latn
|
600-1.0
|
600.0
|
1.0
|
Technology/Computer Science
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 11,
"Qwen/Qwen3-8B": 11,
"bigscience/bloom": 11,
"common-pile/comma-v0.1-1t": 13,
"facebook/xglm-564M": 12,
"google-bert/bert-base-multilingual-cased": 12,
"google/byt5-small": 58,
"google/gemma-2-2b": 11,
"gpt2": 11,
"meta-llama/Llama-3.2-1B": 11,
"microsoft/Phi-3-mini-4k-instruct": 11,
"mistralai/tekken": 11,
"tiktoken/gpt-4o": 11,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
||
The search engine known to prioritize privacy protection is
|
[
"Google Search",
"Perplexity AI",
"Yandex",
"DuckDuckGo"
] | 3
|
D
|
dev
|
Canonical
|
eng_Latn
|
601-1.0
|
601.0
|
1.0
|
Technology/Computer Science
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 9,
"Qwen/Qwen3-8B": 9,
"bigscience/bloom": 10,
"common-pile/comma-v0.1-1t": 11,
"facebook/xglm-564M": 10,
"google-bert/bert-base-multilingual-cased": 12,
"google/byt5-small": 59,
"google/gemma-2-2b": 9,
"gpt2": 9,
"meta-llama/Llama-3.2-1B": 9,
"microsoft/Phi-3-mini-4k-instruct": 12,
"mistralai/tekken": 10,
"tiktoken/gpt-4o": 9,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
||
The professional networking platform most commonly used for job searching is
|
[
"Facebook",
"Twitter",
"LinkedIn",
"Instagram"
] | 2
|
C
|
dev
|
Canonical
|
eng_Latn
|
603-1.0
|
603.0
|
1.0
|
Technology/Computer Science
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 11,
"Qwen/Qwen3-8B": 11,
"bigscience/bloom": 11,
"common-pile/comma-v0.1-1t": 11,
"facebook/xglm-564M": 11,
"google-bert/bert-base-multilingual-cased": 13,
"google/byt5-small": 76,
"google/gemma-2-2b": 11,
"gpt2": 11,
"meta-llama/Llama-3.2-1B": 11,
"microsoft/Phi-3-mini-4k-instruct": 11,
"mistralai/tekken": 11,
"tiktoken/gpt-4o": 11,
"tokenmonster/englishcode-32000-consistent-v1": 10
}
|
||
The programming language popular for data science and machine learning is
|
[
"Java",
"Python",
"C++",
"JavaScript"
] | 1
|
B
|
test
|
Canonical
|
eng_Latn
|
604-1.0
|
604.0
|
1.0
|
Technology/Computer Science
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 11,
"Qwen/Qwen3-8B": 11,
"bigscience/bloom": 11,
"common-pile/comma-v0.1-1t": 13,
"facebook/xglm-564M": 11,
"google-bert/bert-base-multilingual-cased": 11,
"google/byt5-small": 73,
"google/gemma-2-2b": 11,
"gpt2": 11,
"meta-llama/Llama-3.2-1B": 11,
"microsoft/Phi-3-mini-4k-instruct": 11,
"mistralai/tekken": 11,
"tiktoken/gpt-4o": 11,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
||
Root has 2 children. The number of children root has is
|
[
"2",
"1",
"3",
"4"
] | 0
|
A
|
dev
|
Canonical
|
eng_Latn
|
605-1.0
|
605.0
|
1.0
|
Technology/Computer Science
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 13,
"Qwen/Qwen3-8B": 13,
"bigscience/bloom": 12,
"common-pile/comma-v0.1-1t": 14,
"facebook/xglm-564M": 12,
"google-bert/bert-base-multilingual-cased": 12,
"google/byt5-small": 55,
"google/gemma-2-2b": 13,
"gpt2": 12,
"meta-llama/Llama-3.2-1B": 13,
"microsoft/Phi-3-mini-4k-instruct": 13,
"mistralai/tekken": 13,
"tiktoken/gpt-4o": 13,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
||
Start, Process, End. What comes after Start is
|
[
"End",
"Start",
"Middle",
"Process"
] | 3
|
D
|
dev
|
Canonical
|
eng_Latn
|
606-1.0
|
606.0
|
1.0
|
Technology/Computer Science
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 11,
"Qwen/Qwen3-8B": 11,
"bigscience/bloom": 11,
"common-pile/comma-v0.1-1t": 11,
"facebook/xglm-564M": 11,
"google-bert/bert-base-multilingual-cased": 11,
"google/byt5-small": 46,
"google/gemma-2-2b": 11,
"gpt2": 11,
"meta-llama/Llama-3.2-1B": 11,
"microsoft/Phi-3-mini-4k-instruct": 11,
"mistralai/tekken": 11,
"tiktoken/gpt-4o": 11,
"tokenmonster/englishcode-32000-consistent-v1": 10
}
|
||
The letter R in a circuit stands for
|
[
"capacitor",
"resistor",
"inductor",
"battery"
] | 1
|
B
|
dev
|
Canonical
|
eng_Latn
|
607-1.0
|
607.0
|
1.0
|
Technology/Computer Science
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 9,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 36,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
Is JavaScript the same programming language as Java?
|
[
"Yes",
"Sometimes",
"Partially",
"No"
] | 3
|
D
|
dev
|
Canonical
|
eng_Latn
|
609-1.0
|
609.0
|
1.0
|
Technology/Computer Science
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 9,
"Qwen/Qwen3-8B": 9,
"bigscience/bloom": 9,
"common-pile/comma-v0.1-1t": 9,
"facebook/xglm-564M": 9,
"google-bert/bert-base-multilingual-cased": 9,
"google/byt5-small": 52,
"google/gemma-2-2b": 9,
"gpt2": 9,
"meta-llama/Llama-3.2-1B": 9,
"microsoft/Phi-3-mini-4k-instruct": 9,
"mistralai/tekken": 10,
"tiktoken/gpt-4o": 10,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
int(0, 1) represents
|
[
"definite integral from 0 to 1",
"indefinite integral",
"derivative",
"summation"
] | 0
|
A
|
test
|
Canonical
|
eng_Latn
|
Integral ASCII
|
621-1.0
|
621.0
|
1.0
|
Mathematics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 7,
"common-pile/comma-v0.1-1t": 8,
"facebook/xglm-564M": 7,
"google-bert/bert-base-multilingual-cased": 7,
"google/byt5-small": 20,
"google/gemma-2-2b": 8,
"gpt2": 7,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
|
A polygon with three sides is called a
|
[
"square",
"triangle",
"circle",
"rectangle"
] | 1
|
B
|
dev
|
Canonical
|
eng_Latn
|
623-1.0
|
623.0
|
1.0
|
Mathematics/Geometry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 8,
"facebook/xglm-564M": 9,
"google-bert/bert-base-multilingual-cased": 10,
"google/byt5-small": 38,
"google/gemma-2-2b": 8,
"gpt2": 9,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
Tumor protein p53 encodes
|
[
"cell cycle regulator",
"tumor suppressor protein",
"apoptosis factor",
"cancer gene"
] | 1
|
B
|
test
|
Canonical
|
eng_Latn
|
Gene vs protein
|
640-1.0
|
640.0
|
1.0
|
Biology/Medicine
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 7,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 7,
"google/byt5-small": 25,
"google/gemma-2-2b": 6,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 7,
"microsoft/Phi-3-mini-4k-instruct": 9,
"mistralai/tekken": 7,
"tiktoken/gpt-4o": 7,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
|
The process of cell division is
|
[
"Mitosis",
"Meiosis",
"Prophase",
"Anaphase"
] | 0
|
A
|
test
|
Canonical
|
eng_Latn
|
642-1.0
|
642.0
|
1.0
|
Biology
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 31,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 5
}
|
||
The vitamin that prevents scurvy is
|
[
"Vitamin C",
"Vitamin D",
"Vitamin A",
"Vitamin B"
] | 0
|
A
|
test
|
Canonical
|
eng_Latn
|
643-1.0
|
643.0
|
1.0
|
Biology/Medicine
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 8,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 10,
"google/byt5-small": 35,
"google/gemma-2-2b": 7,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 9,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 11
}
|
||
The process of photosynthesis is
|
[
"Plants make food",
"Animals eat plants",
"Cells divide",
"Water evaporates"
] | 0
|
A
|
test
|
Canonical
|
eng_Latn
|
644-1.0
|
644.0
|
1.0
|
Biology
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 5,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 32,
"google/gemma-2-2b": 5,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 5,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
A T G C T A The base that pairs with A is
|
[
"G",
"T",
"C",
"U"
] | 1
|
B
|
dev
|
Canonical
|
eng_Latn
|
645-1.0
|
645.0
|
1.0
|
Biology
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 13,
"Qwen/Qwen3-8B": 13,
"bigscience/bloom": 13,
"common-pile/comma-v0.1-1t": 13,
"facebook/xglm-564M": 14,
"google-bert/bert-base-multilingual-cased": 13,
"google/byt5-small": 41,
"google/gemma-2-2b": 13,
"gpt2": 13,
"meta-llama/Llama-3.2-1B": 13,
"microsoft/Phi-3-mini-4k-instruct": 13,
"mistralai/tekken": 13,
"tiktoken/gpt-4o": 13,
"tokenmonster/englishcode-32000-consistent-v1": 13
}
|
||
Pneumonoultramicroscopicsilicovolcanoconiosis is a type of
|
[
"lung disease",
"eye disease",
"heart disease",
"brain disease"
] | 0
|
A
|
dev
|
Canonical
|
eng_Latn
|
646-1.0
|
646.0
|
1.0
|
Biology/Medicine
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 17,
"Qwen/Qwen3-8B": 21,
"bigscience/bloom": 19,
"common-pile/comma-v0.1-1t": 21,
"facebook/xglm-564M": 17,
"google-bert/bert-base-multilingual-cased": 20,
"google/byt5-small": 58,
"google/gemma-2-2b": 16,
"gpt2": 19,
"meta-llama/Llama-3.2-1B": 19,
"microsoft/Phi-3-mini-4k-instruct": 21,
"mistralai/tekken": 18,
"tiktoken/gpt-4o": 19,
"tokenmonster/englishcode-32000-consistent-v1": 26
}
|
||
Paracetamol is a
|
[
"pain reliever",
"paracentesis",
"paracrine signaling",
"paramagnetic compound"
] | 0
|
A
|
dev
|
Canonical
|
eng_Latn
|
Drug names
|
647-1.0
|
647.0
|
1.0
|
Medicine/Pharmacology
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 5,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 5,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 16,
"google/gemma-2-2b": 4,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 5,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
|
Acetaminophen is a
|
[
"paracentesis",
"paracrine signaling",
"paramagnetic compound",
"pain reliever"
] | 3
|
D
|
dev
|
Canonical
|
eng_Latn
|
648-1.0
|
648.0
|
1.0
|
Medicine/Pharmacology
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 7,
"bigscience/bloom": 7,
"common-pile/comma-v0.1-1t": 5,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 18,
"google/gemma-2-2b": 4,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 7,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
The chemical formula H2SO4 represents
|
[
"hydrogen sulfate",
"sulfur tetroxide",
"bisulfate",
"sulfuric acid"
] | 3
|
D
|
test
|
Canonical
|
eng_Latn
|
660-1.0
|
660.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 8,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 9,
"google/byt5-small": 37,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
The chemical formula CH4 represents
|
[
"methane",
"water",
"benzene",
"oxygen"
] | 0
|
A
|
dev
|
Canonical
|
eng_Latn
|
661-1.0
|
661.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 35,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The chemical formula CO2 represents
|
[
"carbon dioxide",
"carbon monoxide",
"oxygen",
"carbon"
] | 0
|
A
|
test
|
Canonical
|
eng_Latn
|
662-1.0
|
662.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 35,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
The chemical formula O2 represents
|
[
"water",
"oxygen",
"ozone",
"air"
] | 1
|
B
|
test
|
Canonical
|
eng_Latn
|
663-1.0
|
663.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 34,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
The chemical formula H2 represents
|
[
"hydrogen",
"water",
"helium",
"gas"
] | 0
|
A
|
test
|
Canonical
|
eng_Latn
|
664-1.0
|
664.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 34,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
The chemical formula N2 represents
|
[
"ammonia",
"nitrate",
"nitrogen",
"air"
] | 2
|
C
|
test
|
Canonical
|
eng_Latn
|
665-1.0
|
665.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 34,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
The chemical formula NH3 represents
|
[
"nitrogen",
"ammonia",
"hydrogen",
"gas"
] | 1
|
B
|
test
|
Canonical
|
eng_Latn
|
666-1.0
|
666.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 35,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The chemical formula HCl represents
|
[
"sulfuric acid",
"nitric acid",
"hydrochloric acid",
"acetic acid"
] | 2
|
C
|
test
|
Canonical
|
eng_Latn
|
667-1.0
|
667.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 5,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 5,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 35,
"google/gemma-2-2b": 5,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 5,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The chemical formula CaCO3 represents
|
[
"calcium carbonate",
"calcium oxide",
"carbon dioxide",
"magnesium sulfate"
] | 0
|
A
|
test
|
Canonical
|
eng_Latn
|
668-1.0
|
668.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 7,
"Qwen/Qwen3-8B": 7,
"bigscience/bloom": 7,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 7,
"google-bert/bert-base-multilingual-cased": 7,
"google/byt5-small": 37,
"google/gemma-2-2b": 7,
"gpt2": 7,
"meta-llama/Llama-3.2-1B": 7,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 7,
"tiktoken/gpt-4o": 7,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
The chemical formula H2O represents
|
[
"hydrogen",
"oxygen",
"dioxide",
"water"
] | 3
|
D
|
test
|
Canonical
|
eng_Latn
|
669-1.0
|
669.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 7,
"Qwen/Qwen3-8B": 7,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 7,
"google-bert/bert-base-multilingual-cased": 7,
"google/byt5-small": 35,
"google/gemma-2-2b": 7,
"gpt2": 7,
"meta-llama/Llama-3.2-1B": 7,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 7,
"tiktoken/gpt-4o": 7,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The chemical formula for water is
|
[
"H2O",
"H2SO4",
"NaCl",
"CO2"
] | 0
|
A
|
test
|
Canonical
|
eng_Latn
|
670-1.0
|
670.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 33,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
The chemical symbol for gold is
|
[
"Ag",
"Al",
"Ar",
"Au"
] | 3
|
D
|
dev
|
Canonical
|
eng_Latn
|
671-1.0
|
671.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 31,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
The atomic number of carbon is
|
[
"4",
"8",
"6",
"12"
] | 2
|
C
|
test
|
Canonical
|
eng_Latn
|
672-1.0
|
672.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 7,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 30,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 5
}
|
||
The boiling point of water is
|
[
"90°C",
"110°C",
"100°C",
"80°C"
] | 2
|
C
|
test
|
Canonical
|
eng_Latn
|
673-1.0
|
673.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 7,
"google-bert/bert-base-multilingual-cased": 7,
"google/byt5-small": 29,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
The chemical symbol for aluminum is
|
[
"Au",
"Ag",
"Al",
"Fe"
] | 2
|
C
|
dev
|
Canonical
|
eng_Latn
|
674-1.0
|
674.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 7,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 35,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
||
The type of bonds in the carbon dioxide molecule are
|
[
"covalent bonds",
"single bonds",
"triple bonds",
"ionic bonds"
] | 0
|
A
|
dev
|
Unusual formatting
|
eng_Latn
|
Molecular structure
|
675-1.0
|
675.0
|
1.0
|
Chemistry
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 10,
"Qwen/Qwen3-8B": 10,
"bigscience/bloom": 10,
"common-pile/comma-v0.1-1t": 12,
"facebook/xglm-564M": 12,
"google-bert/bert-base-multilingual-cased": 12,
"google/byt5-small": 52,
"google/gemma-2-2b": 10,
"gpt2": 10,
"meta-llama/Llama-3.2-1B": 10,
"microsoft/Phi-3-mini-4k-instruct": 15,
"mistralai/tekken": 10,
"tiktoken/gpt-4o": 10,
"tokenmonster/englishcode-32000-consistent-v1": 12
}
|
|
m/s2 represents
|
[
"velocity",
"distance",
"acceleration",
"speed"
] | 2
|
C
|
dev
|
Canonical
|
eng_Latn
|
Acceleration ASCII
|
680-1.0
|
680.0
|
1.0
|
Physics/Mathematics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 5,
"Qwen/Qwen3-8B": 4,
"bigscience/bloom": 3,
"common-pile/comma-v0.1-1t": 4,
"facebook/xglm-564M": 5,
"google-bert/bert-base-multilingual-cased": 5,
"google/byt5-small": 15,
"google/gemma-2-2b": 5,
"gpt2": 5,
"meta-llama/Llama-3.2-1B": 4,
"microsoft/Phi-3-mini-4k-instruct": 5,
"mistralai/tekken": 4,
"tiktoken/gpt-4o": 4,
"tokenmonster/englishcode-32000-consistent-v1": 4
}
|
|
m3 is the unit of
|
[
"area",
"volume",
"density",
"mass"
] | 1
|
B
|
test
|
Canonical
|
eng_Latn
|
Volume units
|
681-1.0
|
681.0
|
1.0
|
Physics/Mathematics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 17,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 5
}
|
|
m2 is the unit of
|
[
"volume",
"area",
"perimeter",
"density"
] | 1
|
B
|
test
|
Canonical
|
eng_Latn
|
Area units
|
682-1.0
|
682.0
|
1.0
|
Physics/Mathematics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 5,
"google/byt5-small": 17,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 5
}
|
|
kg*m/s2 measures
|
[
"momentum",
"energy",
"force",
"power"
] | 2
|
C
|
dev
|
Canonical
|
eng_Latn
|
Force ASCII
|
683-1.0
|
683.0
|
1.0
|
Physics/Mathematics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 7,
"Qwen/Qwen3-8B": 5,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 5,
"facebook/xglm-564M": 7,
"google-bert/bert-base-multilingual-cased": 7,
"google/byt5-small": 16,
"google/gemma-2-2b": 7,
"gpt2": 7,
"meta-llama/Llama-3.2-1B": 5,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 5,
"tiktoken/gpt-4o": 5,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
|
kg*m2/s2 represents
|
[
"momentum",
"energy",
"force",
"mass"
] | 1
|
B
|
dev
|
Canonical
|
eng_Latn
|
Energy ASCII
|
684-1.0
|
684.0
|
1.0
|
Physics/Mathematics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 7,
"google-bert/bert-base-multilingual-cased": 7,
"google/byt5-small": 19,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
|
Bremsstrahlung is
|
[
"X-ray emission",
"Gamma radiation",
"electromagnetic waves",
"photon radiation"
] | 0
|
A
|
test
|
Canonical
|
eng_Latn
|
690-1.0
|
690.0
|
1.0
|
Physics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 4,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 5,
"google-bert/bert-base-multilingual-cased": 4,
"google/byt5-small": 17,
"google/gemma-2-2b": 4,
"gpt2": 7,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 4,
"tiktoken/gpt-4o": 5,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
||
E=mc2 describes
|
[
"mass-energy equivalence",
"energy conservation",
"momentum transfer",
"force calculation"
] | 0
|
A
|
dev
|
Canonical
|
eng_Latn
|
Energy ASCII
|
691-1.0
|
691.0
|
1.0
|
Physics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 5,
"Qwen/Qwen3-8B": 5,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 5,
"facebook/xglm-564M": 5,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 15,
"google/gemma-2-2b": 5,
"gpt2": 5,
"meta-llama/Llama-3.2-1B": 5,
"microsoft/Phi-3-mini-4k-instruct": 5,
"mistralai/tekken": 5,
"tiktoken/gpt-4o": 5,
"tokenmonster/englishcode-32000-consistent-v1": 5
}
|
|
Bose-Einstein condensation occurs at
|
[
"ultralow temperatures",
"bosonic interactions",
"einstein relativity",
"condensation physics"
] | 0
|
A
|
dev
|
Canonical
|
eng_Latn
|
Quantum phenomenon
|
692-1.0
|
692.0
|
1.0
|
Physics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 7,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 7,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 9,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 36,
"google/gemma-2-2b": 6,
"gpt2": 9,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 11,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
|
The branch of physics that deals with motion is
|
[
"Thermodynamics",
"Quantum Mechanics",
"Classical Mechanics",
"Optics"
] | 2
|
C
|
dev
|
Canonical
|
eng_Latn
|
693-1.0
|
693.0
|
1.0
|
Physics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 9,
"Qwen/Qwen3-8B": 9,
"bigscience/bloom": 9,
"common-pile/comma-v0.1-1t": 10,
"facebook/xglm-564M": 9,
"google-bert/bert-base-multilingual-cased": 9,
"google/byt5-small": 47,
"google/gemma-2-2b": 9,
"gpt2": 9,
"meta-llama/Llama-3.2-1B": 9,
"microsoft/Phi-3-mini-4k-instruct": 10,
"mistralai/tekken": 9,
"tiktoken/gpt-4o": 9,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
||
The theory of relativity was developed by
|
[
"Albert Einstein",
"Isaac Newton",
"Galileo Galilei",
"Stephen Hawking"
] | 0
|
A
|
test
|
Canonical
|
eng_Latn
|
694-1.0
|
694.0
|
1.0
|
Physics
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 7,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 8,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 41,
"google/gemma-2-2b": 7,
"gpt2": 7,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 7,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
Dataset Card for Tokenization Robustness
TokSuite Benchmark (STEM Collection)
Dataset Description
This dataset is the STEM subset of the TokSuite benchmark, designed to evaluate how tokenizer choice affects model behavior under realistic formatting, notation, and surface-form perturbations in technical text. TokSuite includes specialized benchmarks for mathematics and STEM, with the STEM subset containing 44 canonical technical questions paired with a diverse set of targeted perturbations.
- Curated by: R3 Research Team
- License: MIT License
Dataset Summary
TokSuite addresses a fundamental challenge in language model research: understanding how tokenization choices impact model behavior in isolation, independent of architecture, training data, or optimization.
The STEM subset specifically measures model performance on technical and scientific questions under perturbations that commonly arise in real-world STEM communication, including Unicode formatting, mathematical notation, LaTeX representations, spacing changes, and visual styling variants.
Key Features:
- 44 canonical STEM questions covering science, mathematics, engineering, and technical reasoning
- A wide range of formatting- and notation-based perturbations reflecting real-world technical text
- Parallel structure with other TokSuite benchmark subsets
- Designed for controlled robustness evaluation with high baseline accuracy
Supported Tasks
- Multiple-Choice Question Answering: Text completion format with 4 answer choices
- Tokenizer Robustness Evaluation: Measuring performance degradation under surface-form and formatting perturbations
- Technical Text Understanding: Evaluating model robustness on STEM-style content
Languages
The dataset is primarily in English (en), with variations expressed through symbolic, typographic, and Unicode transformations rather than natural-language translation.
Dataset Structure
Data Fields
| Field | Type | Description |
|---|---|---|
question |
string |
The STEM question text |
choices |
list[string] |
4 multiple-choice answer options |
answer |
int64 |
Index of the correct answer |
answer_label |
string |
Letter label of the correct answer |
split |
string |
Dataset split identifier |
subcategories |
string |
Perturbation category |
lang |
string |
Language code (en) |
second_lang |
string |
Optional plain-text or alternative representation |
notes |
string |
Additional context about the perturbation |
id |
string |
Unique question identifier |
set_id |
float64 |
Question set grouping identifier |
variation_id |
float64 |
Variation number within a question set |
vanilla_cos_sim_to_canonical |
dict[string, float] |
Cosine similarity to canonical form (raw tokens) |
trimmed_cos_sim_to_canonical |
dict[string, float] |
Cosine similarity after token normalization |
token_counts |
dict[string, integer] |
Token counts per tokenizer |
Dataset Creation
Curation Rationale
This dataset was created to:
- Systematically evaluate how different tokenization strategies handle STEM-style text
- Measure robustness to formatting, notation, and Unicode-based perturbations
- Isolate tokenizer effects from semantic reasoning difficulty
- Provide standardized benchmarks for technical text robustness analysis
The questions were intentionally designed to be conceptually straightforward, ensuring high canonical accuracy and enabling clean measurement of performance degradation due solely to perturbations.
Source Data
Data Collection and Processing
- Canonical Questions: 44 STEM questions authored in clean, standard technical English
- Perturbations: Each question was transformed using targeted surface-form and formatting variations
- Validation: Model-in-the-loop filtering ensured canonical questions are answerable with high accuracy
Perturbation Categories
Each perturbation preserves the underlying semantic intent of the canonical STEM question while modifying its surface form, notation, or formatting to stress tokenizer behavior. All perturbations are paired with the same canonical question and differ only in representation.
1. Canonical
Clean, standard technical English with conventional notation, spacing, and formatting. This serves as the reference condition for evaluating robustness.
2. Character Deletion
Removes one or more characters from technical terms, symbols, or variables (e.g., markup → markp). These deletions are subtle but often catastrophic for subword tokenization, especially in STEM terminology.
3. Colloquial
Rewrites the question using more informal or descriptive language while preserving technical meaning. This tests robustness to register changes without altering core content.
4. Compounds
Alters compound technical terms by merging or restructuring components (e.g., removing separators or introducing fused forms), changing token boundaries and segmentation behavior.
5. Diacriticized Styling
Introduces decorative or combining diacritics applied to characters in technical text. These perturbations preserve visual similarity but change Unicode code points and normalization behavior.
6. Double-Struck Characters
Replaces standard Latin characters with mathematical double-struck Unicode forms (e.g., R → ℝ, Z → ℤ), commonly used in mathematical notation.
7. Enclosed Characters
Substitutes alphanumeric characters with enclosed Unicode variants (e.g., A → Ⓐ, 1 → ①), which are visually similar but tokenized very differently.
8. Equivalent Expressions
Rewrites the same STEM concept using an alternative but semantically equivalent formulation, such as paraphrasing definitions or reordering explanatory clauses.
9. Fullwidth Characters
Uses fullwidth Unicode forms (e.g., A → A, 1 → 1) instead of standard ASCII characters, altering byte-level and subword tokenization.
10. LaTeX
Represents mathematical expressions or symbols using LaTeX-style notation (e.g., $x^2$, $N_2$, \frac{a}{b}), reflecting common technical writing practices.
11. Morpheme Separation
Artificially splits technical terms into smaller morpheme-like units, increasing sequence length and disrupting learned subword patterns.
12. Scripted Text
Uses scripted or calligraphic Unicode variants of characters (e.g., 𝒜𝒷𝒸) in place of standard Latin letters, stressing visual–semantic mismatch handling.
13. Space Removal
Removes or alters whitespace that is normally meaningful in technical text, such as between variables, units, or multi-word terms.
14. Spelled-Out
Replaces numerals, symbols, or abbreviated technical forms with fully spelled-out textual equivalents (e.g., 2 → two, H2O → water molecule).
15. Strikethrough
Applies strikethrough, combining characters or formatting marks to portions of text, preserving content but introducing visual and Unicode noise.
16. Superscript / Subscript
Uses Unicode superscript and subscript characters (e.g., x², N₂) instead of linear text representations, which often fragment tokenization.
17. Typographical Errors
Introduces realistic typos such as missing letters, duplicated characters, or minor corruptions common in fast technical writing.
18. Unicode Formatting
Applies Unicode formatting characters that affect text rendering or directionality while leaving the visible content largely unchanged.
19. Unusual Formatting
Introduces nonstandard layout, punctuation, or visual formatting patterns that are uncommon but realistic in technical documents.
20. Upside-Down / Rotated
Uses visually rotated or upside-down Unicode characters that resemble standard characters but differ at the code-point level.
Considerations for Using the Data
Social Impact of Dataset
This dataset supports the development of more robust language models for technical and scientific domains, improving reliability in education, research, and engineering applications.
Discussion of Biases
- Domain focus: Emphasizes STEM-style technical text rather than natural language discourse
- Formatting-centric: Perturbations focus on surface form, not conceptual difficulty
- English-centric: Uses English technical text, though many perturbations are language-agnostic
- Question simplicity: Designed for robustness evaluation rather than deep problem-solving
Other Known Limitations
- Evaluation-only dataset (no training split)
- Multiple-choice format
- Limited question count per perturbation
- Results may differ for long-form or open-ended STEM reasoning
Additional Information
Dataset Curators
The dataset was curated by the TokSuite research team at R3.
Licensing Information
MIT License
Citation Information
If you use this dataset in your research, please cite the TokSuite paper:
@inproceedings{toksuite2026, title={TokSuite: Measuring the Impact of Tokenizer Choice on Language Model Behavior}, author={Altıntaş, Gül Sena and Ehghaghi, Malikeh and Lester, Brian and Liu, Fengyuan and Zhao, Wanru and Ciccone, Marco and Raffel, Colin}, booktitle={Preprint}, year={2026}, arxiv={https://arxiv.org/abs/2512.20757}, url={TBD} }
Paper: TokSuite: Measuring the Impact of Tokenizer Choice on Language Model Behavior
Contributions
This dataset is part of TokSuite, which includes:
- 14 language models with identical architectures but different tokenizers
- Multilingual benchmark datasets (English, Turkish, Italian, Farsi, Chinese)
- Comprehensive analysis of tokenization's impact on model behavior
Contact
For questions or issues related to this dataset, please refer to the TokSuite project or contact the authors of the paper.
Part of the TokSuite Project
Understanding Tokenization's Role in Language Model Behavior
- Downloads last month
- 160