Dataset Viewer
Auto-converted to Parquet Duplicate
question
stringlengths
14
72
choices
listlengths
4
4
answer
int64
0
3
answer_label
stringclasses
4 values
split
stringclasses
1 value
subcategories
stringclasses
2 values
category
stringclasses
2 values
lang
stringclasses
1 value
second_lang
stringclasses
1 value
notes
stringclasses
2 values
id
stringlengths
7
7
set_id
stringlengths
5
5
variation_id
stringclasses
7 values
vanilla_cos_sim_to_canonical
dict
trimmed_cos_sim_to_canonical
dict
token_counts
dict
Dr Smith is a doctor. Occupation of Dr Smith is
[ "teach", "judge", "law", "doc" ]
3
D
test
Abbreviations
Script / Orthography
eng_Latn
300-0.2
300.0
0.2
{ "CohereLabs/aya-expanse-8b": 1, "Qwen/Qwen3-8B": 1, "bigscience/bloom": 1.0000001192092896, "common-pile/comma-v0.1-1t": 0.9999998211860657, "facebook/xglm-564M": 1.0000001192092896, "google-bert/bert-base-multilingual-cased": 0.9999998211860657, "google/byt5-small": 1.0000001192092896, "google/gemma-2-2b": 0.9999998807907104, "gpt2": 0.9999998211860657, "meta-llama/Llama-3.2-1B": 1.0000001192092896, "microsoft/Phi-3-mini-4k-instruct": 0.9999998807907104, "mistralai/tekken": 0.9999998807907104, "tiktoken/gpt-4o": 0.9999999403953552, "tokenmonster/englishcode-32000-consistent-v1": 1.0000001192092896 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687 }
{ "CohereLabs/aya-expanse-8b": 11, "Qwen/Qwen3-8B": 11, "bigscience/bloom": 12, "common-pile/comma-v0.1-1t": 13, "facebook/xglm-564M": 13, "google-bert/bert-base-multilingual-cased": 14, "google/byt5-small": 47, "google/gemma-2-2b": 11, "gpt2": 12, "meta-llama/Llama-3.2-1B": 11, "microsoft/Phi-3-mini-4k-instruct": 13, "mistralai/tekken": 11, "tiktoken/gpt-4o": 12, "tokenmonster/englishcode-32000-consistent-v1": 11 }
Dr Smith is a doctor. Occupation of Dr Smith is
[ "Prof", "MD", "Hon", "Esq" ]
1
B
test
Abbreviations
Script / Orthography
eng_Latn
300-0.3
300.0
0.3
{ "CohereLabs/aya-expanse-8b": 1, "Qwen/Qwen3-8B": 1, "bigscience/bloom": 1.0000001192092896, "common-pile/comma-v0.1-1t": 0.9999998211860657, "facebook/xglm-564M": 1.0000001192092896, "google-bert/bert-base-multilingual-cased": 0.9999998211860657, "google/byt5-small": 1.0000001192092896, "google/gemma-2-2b": 0.9999998807907104, "gpt2": 0.9999998211860657, "meta-llama/Llama-3.2-1B": 1.0000001192092896, "microsoft/Phi-3-mini-4k-instruct": 0.9999998807907104, "mistralai/tekken": 0.9999998807907104, "tiktoken/gpt-4o": 0.9999999403953552, "tokenmonster/englishcode-32000-consistent-v1": 1.0000001192092896 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687 }
{ "CohereLabs/aya-expanse-8b": 11, "Qwen/Qwen3-8B": 11, "bigscience/bloom": 12, "common-pile/comma-v0.1-1t": 13, "facebook/xglm-564M": 13, "google-bert/bert-base-multilingual-cased": 14, "google/byt5-small": 47, "google/gemma-2-2b": 11, "gpt2": 12, "meta-llama/Llama-3.2-1B": 11, "microsoft/Phi-3-mini-4k-instruct": 13, "mistralai/tekken": 11, "tiktoken/gpt-4o": 12, "tokenmonster/englishcode-32000-consistent-v1": 11 }
Dr Smith is an MD. Occipation of Dr Smith is
[ "doctor", "teacher", "judge", "lawyer" ]
0
A
test
Abbreviations
Script / Orthography
eng_Latn
300-0.4
300.0
0.4
{ "CohereLabs/aya-expanse-8b": 0.8465678691864014, "Qwen/Qwen3-8B": 0.8233134150505066, "bigscience/bloom": 0.873887836933136, "common-pile/comma-v0.1-1t": 0.7490841150283813, "facebook/xglm-564M": 0.8408849239349365, "google-bert/bert-base-multilingual-cased": 0.7985623478889465, "google/byt5-small": 0.9447391033172607, "google/gemma-2-2b": 0.835648775100708, "gpt2": 0.8251903057098389, "meta-llama/Llama-3.2-1B": 0.8439704775810242, "microsoft/Phi-3-mini-4k-instruct": 0.8045406937599182, "mistralai/tekken": 0.8300552368164062, "tiktoken/gpt-4o": 0.8318107724189758, "tokenmonster/englishcode-32000-consistent-v1": 0.7099841237068176 }
{ "CohereLabs/aya-expanse-8b": 0.37533700466156006, "Qwen/Qwen3-8B": 0.23539593815803528, "bigscience/bloom": 0.4065379500389099, "common-pile/comma-v0.1-1t": 0.13768307864665985, "facebook/xglm-564M": 0.3379814028739929, "google-bert/bert-base-multilingual-cased": 0.4463389217853546, "google/byt5-small": 0.6483171582221985, "google/gemma-2-2b": 0.35768020153045654, "gpt2": 0.3287959694862366, "meta-llama/Llama-3.2-1B": 0.3190908432006836, "microsoft/Phi-3-mini-4k-instruct": 0.2861732244491577, "mistralai/tekken": 0.27699482440948486, "tiktoken/gpt-4o": 0.3433341383934021, "tokenmonster/englishcode-32000-consistent-v1": 0.19402597844600677 }
{ "CohereLabs/aya-expanse-8b": 12, "Qwen/Qwen3-8B": 12, "bigscience/bloom": 13, "common-pile/comma-v0.1-1t": 15, "facebook/xglm-564M": 13, "google-bert/bert-base-multilingual-cased": 14, "google/byt5-small": 44, "google/gemma-2-2b": 12, "gpt2": 12, "meta-llama/Llama-3.2-1B": 12, "microsoft/Phi-3-mini-4k-instruct": 14, "mistralai/tekken": 12, "tiktoken/gpt-4o": 12, "tokenmonster/englishcode-32000-consistent-v1": 16 }
The price of this house is 1.03M dollars The cost of this house is
[ "1,030,000 dollars", "1.030.000 dollars", "1,030,000 dollars", "1,030.000 dollars" ]
0
A
test
Abbreviations, Unit combinations
Script / Orthography, Mathematical & Scientific Notation
eng_Latn
changed options to accommodate M
302-0.3
302.0
0.3
{ "CohereLabs/aya-expanse-8b": 0.9055666327476501, "Qwen/Qwen3-8B": 0.9031217694282532, "bigscience/bloom": 0.894109845161438, "common-pile/comma-v0.1-1t": 0.9072467088699341, "facebook/xglm-564M": 0.875411868095398, "google-bert/bert-base-multilingual-cased": 0.8923179507255554, "google/byt5-small": 0.9754881858825684, "google/gemma-2-2b": 0.9033129215240479, "gpt2": 0.8858630061149597, "meta-llama/Llama-3.2-1B": 0.8717429041862488, "microsoft/Phi-3-mini-4k-instruct": 0.8849417567253113, "mistralai/tekken": 0.881121039390564, "tiktoken/gpt-4o": 0.8818330764770508, "tokenmonster/englishcode-32000-consistent-v1": 0.8387635350227356 }
{ "CohereLabs/aya-expanse-8b": 0.5830617547035217, "Qwen/Qwen3-8B": 0.55271977186203, "bigscience/bloom": 0.4354766607284546, "common-pile/comma-v0.1-1t": 0.40262073278427124, "facebook/xglm-564M": 0.28853747248649597, "google-bert/bert-base-multilingual-cased": 0.42490333318710327, "google/byt5-small": 0.6922744512557983, "google/gemma-2-2b": 0.5299580097198486, "gpt2": 0.3897722065448761, "meta-llama/Llama-3.2-1B": 0.3892613649368286, "microsoft/Phi-3-mini-4k-instruct": 0.5016849637031555, "mistralai/tekken": 0.5062558054924011, "tiktoken/gpt-4o": 0.37294676899909973, "tokenmonster/englishcode-32000-consistent-v1": 0.31927287578582764 }
{ "CohereLabs/aya-expanse-8b": 19, "Qwen/Qwen3-8B": 19, "bigscience/bloom": 17, "common-pile/comma-v0.1-1t": 20, "facebook/xglm-564M": 16, "google-bert/bert-base-multilingual-cased": 17, "google/byt5-small": 66, "google/gemma-2-2b": 19, "gpt2": 17, "meta-llama/Llama-3.2-1B": 18, "microsoft/Phi-3-mini-4k-instruct": 19, "mistralai/tekken": 19, "tiktoken/gpt-4o": 18, "tokenmonster/englishcode-32000-consistent-v1": 14 }
The # of continents on Earth is
[ "7", "5", "6", "8" ]
0
A
test
Abbreviations
Script / Orthography
eng_Latn
304-0.7
304.0
0.7
{ "CohereLabs/aya-expanse-8b": 0.8543572425842285, "Qwen/Qwen3-8B": 0.8774767518043518, "bigscience/bloom": 0.8670704960823059, "common-pile/comma-v0.1-1t": 0.8822222352027893, "facebook/xglm-564M": 0.8440374135971069, "google-bert/bert-base-multilingual-cased": 0.8720088005065918, "google/byt5-small": 0.9478123188018799, "google/gemma-2-2b": 0.8665192127227783, "gpt2": 0.8531822562217712, "meta-llama/Llama-3.2-1B": 0.8680265545845032, "microsoft/Phi-3-mini-4k-instruct": 0.8711134791374207, "mistralai/tekken": 0.8590569496154785, "tiktoken/gpt-4o": 0.8604445457458496, "tokenmonster/englishcode-32000-consistent-v1": 0.7292405962944031 }
{ "CohereLabs/aya-expanse-8b": 0.15336325764656067, "Qwen/Qwen3-8B": 0.17457246780395508, "bigscience/bloom": 0.18650875985622406, "common-pile/comma-v0.1-1t": 0.2236202359199524, "facebook/xglm-564M": 0.05189888924360275, "google-bert/bert-base-multilingual-cased": 0.11308808624744415, "google/byt5-small": -0.02643284946680069, "google/gemma-2-2b": 0.173978790640831, "gpt2": 0.15644428133964539, "meta-llama/Llama-3.2-1B": 0.19826795160770416, "microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529, "mistralai/tekken": 0.16212503612041473, "tiktoken/gpt-4o": 0.1929726004600525, "tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649 }
{ "CohereLabs/aya-expanse-8b": 7, "Qwen/Qwen3-8B": 7, "bigscience/bloom": 7, "common-pile/comma-v0.1-1t": 8, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 31, "google/gemma-2-2b": 7, "gpt2": 7, "meta-llama/Llama-3.2-1B": 7, "microsoft/Phi-3-mini-4k-instruct": 8, "mistralai/tekken": 7, "tiktoken/gpt-4o": 7, "tokenmonster/englishcode-32000-consistent-v1": 7 }
The no. of continents on Earth is
[ "5", "6", "8", "7" ]
3
D
test
Abbreviations
Script / Orthography
eng_Latn
304-0.8
304.0
0.8
{ "CohereLabs/aya-expanse-8b": 0.7867254614830017, "Qwen/Qwen3-8B": 0.824273943901062, "bigscience/bloom": 0.811360776424408, "common-pile/comma-v0.1-1t": 0.8064047694206238, "facebook/xglm-564M": 0.8018571138381958, "google-bert/bert-base-multilingual-cased": 0.8182386159896851, "google/byt5-small": 0.9467005133628845, "google/gemma-2-2b": 0.804913341999054, "gpt2": 0.7843954563140869, "meta-llama/Llama-3.2-1B": 0.7866487503051758, "microsoft/Phi-3-mini-4k-instruct": 0.8095418810844421, "mistralai/tekken": 0.7949608564376831, "tiktoken/gpt-4o": 0.7873102426528931, "tokenmonster/englishcode-32000-consistent-v1": 0.7079726457595825 }
{ "CohereLabs/aya-expanse-8b": 0.07414855808019638, "Qwen/Qwen3-8B": 0.03043564409017563, "bigscience/bloom": 0.07808922976255417, "common-pile/comma-v0.1-1t": -0.02923409827053547, "facebook/xglm-564M": 0.06573980301618576, "google-bert/bert-base-multilingual-cased": -0.015136616304516792, "google/byt5-small": -0.020509906113147736, "google/gemma-2-2b": 0.0652402713894844, "gpt2": 0.008664418943226337, "meta-llama/Llama-3.2-1B": -0.008986718952655792, "microsoft/Phi-3-mini-4k-instruct": -0.02389761433005333, "mistralai/tekken": 0.06866628676652908, "tiktoken/gpt-4o": 0.03231961280107498, "tokenmonster/englishcode-32000-consistent-v1": 0.13605159521102905 }
{ "CohereLabs/aya-expanse-8b": 8, "Qwen/Qwen3-8B": 8, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 9, "facebook/xglm-564M": 9, "google-bert/bert-base-multilingual-cased": 9, "google/byt5-small": 33, "google/gemma-2-2b": 8, "gpt2": 8, "meta-llama/Llama-3.2-1B": 8, "microsoft/Phi-3-mini-4k-instruct": 9, "mistralai/tekken": 8, "tiktoken/gpt-4o": 8, "tokenmonster/englishcode-32000-consistent-v1": 7 }
The capital city of IR is
[ "Mashhad", "Baghdad", "Tehran", "Isfahan" ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
305-0.6
305.0
0.6
{ "CohereLabs/aya-expanse-8b": 0.8646732568740845, "Qwen/Qwen3-8B": 0.8861888647079468, "bigscience/bloom": 0.8826625347137451, "common-pile/comma-v0.1-1t": 0.8774277567863464, "facebook/xglm-564M": 0.868166446685791, "google-bert/bert-base-multilingual-cased": 0.8571589589118958, "google/byt5-small": 0.9628745317459106, "google/gemma-2-2b": 0.875096321105957, "gpt2": 0.8794789910316467, "meta-llama/Llama-3.2-1B": 0.8609308004379272, "microsoft/Phi-3-mini-4k-instruct": 0.864482581615448, "mistralai/tekken": 0.851871907711029, "tiktoken/gpt-4o": 0.8562781810760498, "tokenmonster/englishcode-32000-consistent-v1": 0.7831379771232605 }
{ "CohereLabs/aya-expanse-8b": 0.1503104269504547, "Qwen/Qwen3-8B": 0.22655273973941803, "bigscience/bloom": 0.24181559681892395, "common-pile/comma-v0.1-1t": 0.22090619802474976, "facebook/xglm-564M": 0.1887182891368866, "google-bert/bert-base-multilingual-cased": 0.15784157812595367, "google/byt5-small": 0.24735668301582336, "google/gemma-2-2b": 0.18847647309303284, "gpt2": 0.25452691316604614, "meta-llama/Llama-3.2-1B": 0.16874276101589203, "microsoft/Phi-3-mini-4k-instruct": 0.13977181911468506, "mistralai/tekken": 0.13231658935546875, "tiktoken/gpt-4o": 0.1192522943019867, "tokenmonster/englishcode-32000-consistent-v1": 0.10740450024604797 }
{ "CohereLabs/aya-expanse-8b": 6, "Qwen/Qwen3-8B": 6, "bigscience/bloom": 6, "common-pile/comma-v0.1-1t": 7, "facebook/xglm-564M": 6, "google-bert/bert-base-multilingual-cased": 6, "google/byt5-small": 25, "google/gemma-2-2b": 6, "gpt2": 6, "meta-llama/Llama-3.2-1B": 6, "microsoft/Phi-3-mini-4k-instruct": 6, "mistralai/tekken": 6, "tiktoken/gpt-4o": 6, "tokenmonster/englishcode-32000-consistent-v1": 6 }
The # of days in a week is
[ "7", "5", "6", "8" ]
0
A
test
Abbreviations
Script / Orthography
eng_Latn
306-0.5
306.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.8769029378890991, "Qwen/Qwen3-8B": 0.8977583050727844, "bigscience/bloom": 0.8890756368637085, "common-pile/comma-v0.1-1t": 0.9004440307617188, "facebook/xglm-564M": 0.8400768041610718, "google-bert/bert-base-multilingual-cased": 0.8637771010398865, "google/byt5-small": 0.9125043749809265, "google/gemma-2-2b": 0.8880384564399719, "gpt2": 0.8734878897666931, "meta-llama/Llama-3.2-1B": 0.8814516067504883, "microsoft/Phi-3-mini-4k-instruct": 0.8667116165161133, "mistralai/tekken": 0.8863945007324219, "tiktoken/gpt-4o": 0.8898459672927856, "tokenmonster/englishcode-32000-consistent-v1": 0.758175253868103 }
{ "CohereLabs/aya-expanse-8b": 0.15336325764656067, "Qwen/Qwen3-8B": 0.17457246780395508, "bigscience/bloom": 0.18650875985622406, "common-pile/comma-v0.1-1t": 0.2236202359199524, "facebook/xglm-564M": 0.05189888924360275, "google-bert/bert-base-multilingual-cased": 0.11308808624744415, "google/byt5-small": -0.02643284946680069, "google/gemma-2-2b": 0.173978790640831, "gpt2": 0.15644428133964539, "meta-llama/Llama-3.2-1B": 0.19826795160770416, "microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529, "mistralai/tekken": 0.16212503612041473, "tiktoken/gpt-4o": 0.1929726004600525, "tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649 }
{ "CohereLabs/aya-expanse-8b": 8, "Qwen/Qwen3-8B": 8, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 10, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 26, "google/gemma-2-2b": 8, "gpt2": 8, "meta-llama/Llama-3.2-1B": 8, "microsoft/Phi-3-mini-4k-instruct": 8, "mistralai/tekken": 8, "tiktoken/gpt-4o": 8, "tokenmonster/englishcode-32000-consistent-v1": 7 }
The # of hours in a day is
[ "24", "20", "25", "30" ]
0
A
test
Abbreviations
Script / Orthography
eng_Latn
307-0.5
307.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.8725788593292236, "Qwen/Qwen3-8B": 0.8972941040992737, "bigscience/bloom": 0.8836281299591064, "common-pile/comma-v0.1-1t": 0.8991660475730896, "facebook/xglm-564M": 0.8364160656929016, "google-bert/bert-base-multilingual-cased": 0.862697422504425, "google/byt5-small": 0.9182883501052856, "google/gemma-2-2b": 0.8852136135101318, "gpt2": 0.8709038496017456, "meta-llama/Llama-3.2-1B": 0.8853892683982849, "microsoft/Phi-3-mini-4k-instruct": 0.8643823862075806, "mistralai/tekken": 0.8764235377311707, "tiktoken/gpt-4o": 0.8882663249969482, "tokenmonster/englishcode-32000-consistent-v1": 0.7297150492668152 }
{ "CohereLabs/aya-expanse-8b": 0.15336325764656067, "Qwen/Qwen3-8B": 0.17457246780395508, "bigscience/bloom": 0.18650875985622406, "common-pile/comma-v0.1-1t": 0.2236202359199524, "facebook/xglm-564M": 0.05189888924360275, "google-bert/bert-base-multilingual-cased": 0.11308808624744415, "google/byt5-small": -0.02643284946680069, "google/gemma-2-2b": 0.173978790640831, "gpt2": 0.15644428133964539, "meta-llama/Llama-3.2-1B": 0.19826795160770416, "microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529, "mistralai/tekken": 0.16212503612041473, "tiktoken/gpt-4o": 0.1929726004600525, "tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649 }
{ "CohereLabs/aya-expanse-8b": 8, "Qwen/Qwen3-8B": 8, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 10, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 26, "google/gemma-2-2b": 8, "gpt2": 8, "meta-llama/Llama-3.2-1B": 8, "microsoft/Phi-3-mini-4k-instruct": 8, "mistralai/tekken": 8, "tiktoken/gpt-4o": 8, "tokenmonster/englishcode-32000-consistent-v1": 7 }
The # of legs a cow has is
[ "8", "4", "3", "5" ]
1
B
test
Abbreviations
Script / Orthography
eng_Latn
308-0.5
308.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.8784412145614624, "Qwen/Qwen3-8B": 0.903361439704895, "bigscience/bloom": 0.8900041580200195, "common-pile/comma-v0.1-1t": 0.8913084268569946, "facebook/xglm-564M": 0.8609603047370911, "google-bert/bert-base-multilingual-cased": 0.8892489671707153, "google/byt5-small": 0.9177855849266052, "google/gemma-2-2b": 0.8813795447349548, "gpt2": 0.877772331237793, "meta-llama/Llama-3.2-1B": 0.8910291790962219, "microsoft/Phi-3-mini-4k-instruct": 0.8654680848121643, "mistralai/tekken": 0.8807762265205383, "tiktoken/gpt-4o": 0.8883159160614014, "tokenmonster/englishcode-32000-consistent-v1": 0.7710891962051392 }
{ "CohereLabs/aya-expanse-8b": 0.15336325764656067, "Qwen/Qwen3-8B": 0.17457246780395508, "bigscience/bloom": 0.18650875985622406, "common-pile/comma-v0.1-1t": 0.2236202359199524, "facebook/xglm-564M": 0.05189888924360275, "google-bert/bert-base-multilingual-cased": 0.11308808624744415, "google/byt5-small": -0.02643284946680069, "google/gemma-2-2b": 0.173978790640831, "gpt2": 0.15644428133964539, "meta-llama/Llama-3.2-1B": 0.19826795160770416, "microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529, "mistralai/tekken": 0.16212503612041473, "tiktoken/gpt-4o": 0.1929726004600525, "tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649 }
{ "CohereLabs/aya-expanse-8b": 8, "Qwen/Qwen3-8B": 8, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 9, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 9, "google/byt5-small": 26, "google/gemma-2-2b": 8, "gpt2": 8, "meta-llama/Llama-3.2-1B": 8, "microsoft/Phi-3-mini-4k-instruct": 8, "mistralai/tekken": 8, "tiktoken/gpt-4o": 8, "tokenmonster/englishcode-32000-consistent-v1": 8 }
The # of minutes in 2 hours is
[ "100", "140", "120", "90" ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
309-0.5
309.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.8936583995819092, "Qwen/Qwen3-8B": 0.9028605222702026, "bigscience/bloom": 0.901308000087738, "common-pile/comma-v0.1-1t": 0.9184684753417969, "facebook/xglm-564M": 0.8589028716087341, "google-bert/bert-base-multilingual-cased": 0.8760905265808105, "google/byt5-small": 0.9449029564857483, "google/gemma-2-2b": 0.9074434041976929, "gpt2": 0.8900853991508484, "meta-llama/Llama-3.2-1B": 0.9088870286941528, "microsoft/Phi-3-mini-4k-instruct": 0.8868902325630188, "mistralai/tekken": 0.9065263271331787, "tiktoken/gpt-4o": 0.9071903824806213, "tokenmonster/englishcode-32000-consistent-v1": 0.7755128145217896 }
{ "CohereLabs/aya-expanse-8b": 0.15336325764656067, "Qwen/Qwen3-8B": 0.17457246780395508, "bigscience/bloom": 0.18650875985622406, "common-pile/comma-v0.1-1t": 0.2236202359199524, "facebook/xglm-564M": 0.05189888924360275, "google-bert/bert-base-multilingual-cased": 0.11308808624744415, "google/byt5-small": -0.02643284946680069, "google/gemma-2-2b": 0.173978790640831, "gpt2": 0.15644428133964539, "meta-llama/Llama-3.2-1B": 0.19826795160770416, "microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529, "mistralai/tekken": 0.16212503612041473, "tiktoken/gpt-4o": 0.1929726004600525, "tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649 }
{ "CohereLabs/aya-expanse-8b": 9, "Qwen/Qwen3-8B": 9, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 11, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 30, "google/gemma-2-2b": 9, "gpt2": 8, "meta-llama/Llama-3.2-1B": 9, "microsoft/Phi-3-mini-4k-instruct": 9, "mistralai/tekken": 9, "tiktoken/gpt-4o": 9, "tokenmonster/englishcode-32000-consistent-v1": 8 }
The # of months in a year is
[ "10", "11", "12", "13" ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
310-0.5
310.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.8738687038421631, "Qwen/Qwen3-8B": 0.8943588733673096, "bigscience/bloom": 0.8882232308387756, "common-pile/comma-v0.1-1t": 0.9044252634048462, "facebook/xglm-564M": 0.8348524570465088, "google-bert/bert-base-multilingual-cased": 0.8641276359558105, "google/byt5-small": 0.9306106567382812, "google/gemma-2-2b": 0.8897289633750916, "gpt2": 0.8724748492240906, "meta-llama/Llama-3.2-1B": 0.8823840022087097, "microsoft/Phi-3-mini-4k-instruct": 0.8683964610099792, "mistralai/tekken": 0.8803357481956482, "tiktoken/gpt-4o": 0.8929274678230286, "tokenmonster/englishcode-32000-consistent-v1": 0.7467564344406128 }
{ "CohereLabs/aya-expanse-8b": 0.15336325764656067, "Qwen/Qwen3-8B": 0.17457246780395508, "bigscience/bloom": 0.18650875985622406, "common-pile/comma-v0.1-1t": 0.2236202359199524, "facebook/xglm-564M": 0.05189888924360275, "google-bert/bert-base-multilingual-cased": 0.11308808624744415, "google/byt5-small": -0.02643284946680069, "google/gemma-2-2b": 0.173978790640831, "gpt2": 0.15644428133964539, "meta-llama/Llama-3.2-1B": 0.19826795160770416, "microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529, "mistralai/tekken": 0.16212503612041473, "tiktoken/gpt-4o": 0.1929726004600525, "tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649 }
{ "CohereLabs/aya-expanse-8b": 8, "Qwen/Qwen3-8B": 8, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 10, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 28, "google/gemma-2-2b": 8, "gpt2": 8, "meta-llama/Llama-3.2-1B": 8, "microsoft/Phi-3-mini-4k-instruct": 8, "mistralai/tekken": 8, "tiktoken/gpt-4o": 8, "tokenmonster/englishcode-32000-consistent-v1": 7 }
The # of seconds in a minute is
[ "60", "50", "100", "30" ]
0
A
test
Abbreviations
Script / Orthography
eng_Latn
311-0.5
311.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.8620327711105347, "Qwen/Qwen3-8B": 0.8879849910736084, "bigscience/bloom": 0.8815665245056152, "common-pile/comma-v0.1-1t": 0.8988855481147766, "facebook/xglm-564M": 0.825326144695282, "google-bert/bert-base-multilingual-cased": 0.8537793159484863, "google/byt5-small": 0.9459080100059509, "google/gemma-2-2b": 0.8843560218811035, "gpt2": 0.8683614134788513, "meta-llama/Llama-3.2-1B": 0.8789793848991394, "microsoft/Phi-3-mini-4k-instruct": 0.8600611090660095, "mistralai/tekken": 0.8747711777687073, "tiktoken/gpt-4o": 0.8857784271240234, "tokenmonster/englishcode-32000-consistent-v1": 0.7317348122596741 }
{ "CohereLabs/aya-expanse-8b": 0.15336325764656067, "Qwen/Qwen3-8B": 0.17457246780395508, "bigscience/bloom": 0.18650875985622406, "common-pile/comma-v0.1-1t": 0.2236202359199524, "facebook/xglm-564M": 0.05189888924360275, "google-bert/bert-base-multilingual-cased": 0.11308808624744415, "google/byt5-small": -0.02643284946680069, "google/gemma-2-2b": 0.173978790640831, "gpt2": 0.15644428133964539, "meta-llama/Llama-3.2-1B": 0.19826795160770416, "microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529, "mistralai/tekken": 0.16212503612041473, "tiktoken/gpt-4o": 0.1929726004600525, "tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649 }
{ "CohereLabs/aya-expanse-8b": 8, "Qwen/Qwen3-8B": 8, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 10, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 31, "google/gemma-2-2b": 8, "gpt2": 8, "meta-llama/Llama-3.2-1B": 8, "microsoft/Phi-3-mini-4k-instruct": 8, "mistralai/tekken": 8, "tiktoken/gpt-4o": 8, "tokenmonster/englishcode-32000-consistent-v1": 7 }
The # of sides a hexagon has is
[ "5", "7", "6", "8" ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
312-0.5
312.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.862587571144104, "Qwen/Qwen3-8B": 0.9136554002761841, "bigscience/bloom": 0.9019626975059509, "common-pile/comma-v0.1-1t": 0.897915244102478, "facebook/xglm-564M": 0.904183030128479, "google-bert/bert-base-multilingual-cased": 0.8995038270950317, "google/byt5-small": 0.9488453269004822, "google/gemma-2-2b": 0.8825085163116455, "gpt2": 0.8939532041549683, "meta-llama/Llama-3.2-1B": 0.9000940322875977, "microsoft/Phi-3-mini-4k-instruct": 0.8842352628707886, "mistralai/tekken": 0.8938607573509216, "tiktoken/gpt-4o": 0.9065819382667542, "tokenmonster/englishcode-32000-consistent-v1": 0.8501964807510376 }
{ "CohereLabs/aya-expanse-8b": 0.15336325764656067, "Qwen/Qwen3-8B": 0.17457246780395508, "bigscience/bloom": 0.18650875985622406, "common-pile/comma-v0.1-1t": 0.2236202359199524, "facebook/xglm-564M": 0.05189888924360275, "google-bert/bert-base-multilingual-cased": 0.11308808624744415, "google/byt5-small": -0.02643284946680069, "google/gemma-2-2b": 0.173978790640831, "gpt2": 0.15644428133964539, "meta-llama/Llama-3.2-1B": 0.19826795160770416, "microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529, "mistralai/tekken": 0.16212503612041473, "tiktoken/gpt-4o": 0.1929726004600525, "tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649 }
{ "CohereLabs/aya-expanse-8b": 8, "Qwen/Qwen3-8B": 9, "bigscience/bloom": 9, "common-pile/comma-v0.1-1t": 10, "facebook/xglm-564M": 10, "google-bert/bert-base-multilingual-cased": 10, "google/byt5-small": 31, "google/gemma-2-2b": 8, "gpt2": 9, "meta-llama/Llama-3.2-1B": 9, "microsoft/Phi-3-mini-4k-instruct": 9, "mistralai/tekken": 9, "tiktoken/gpt-4o": 9, "tokenmonster/englishcode-32000-consistent-v1": 11 }
The # of sides a triangle has is
[ "2", "4", "3", "5" ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
313-0.4
313.0
0.4
{ "CohereLabs/aya-expanse-8b": 0.8708062171936035, "Qwen/Qwen3-8B": 0.9027498364448547, "bigscience/bloom": 0.8900589942932129, "common-pile/comma-v0.1-1t": 0.8905171155929565, "facebook/xglm-564M": 0.8620559573173523, "google-bert/bert-base-multilingual-cased": 0.8632290959358215, "google/byt5-small": 0.9473434090614319, "google/gemma-2-2b": 0.8901822566986084, "gpt2": 0.8811707496643066, "meta-llama/Llama-3.2-1B": 0.8887614607810974, "microsoft/Phi-3-mini-4k-instruct": 0.8653461933135986, "mistralai/tekken": 0.8795601725578308, "tiktoken/gpt-4o": 0.8902370929718018, "tokenmonster/englishcode-32000-consistent-v1": 0.7794355750083923 }
{ "CohereLabs/aya-expanse-8b": 0.15336325764656067, "Qwen/Qwen3-8B": 0.17457246780395508, "bigscience/bloom": 0.18650875985622406, "common-pile/comma-v0.1-1t": 0.2236202359199524, "facebook/xglm-564M": 0.05189888924360275, "google-bert/bert-base-multilingual-cased": 0.11308808624744415, "google/byt5-small": -0.02643284946680069, "google/gemma-2-2b": 0.173978790640831, "gpt2": 0.15644428133964539, "meta-llama/Llama-3.2-1B": 0.19826795160770416, "microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529, "mistralai/tekken": 0.16212503612041473, "tiktoken/gpt-4o": 0.1929726004600525, "tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649 }
{ "CohereLabs/aya-expanse-8b": 8, "Qwen/Qwen3-8B": 8, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 9, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 32, "google/gemma-2-2b": 8, "gpt2": 8, "meta-llama/Llama-3.2-1B": 8, "microsoft/Phi-3-mini-4k-instruct": 8, "mistralai/tekken": 8, "tiktoken/gpt-4o": 8, "tokenmonster/englishcode-32000-consistent-v1": 8 }
In "I work at Apple", Apple is a
[ "pers.", "cty.", "fr.", "co." ]
3
D
test
Abbreviations
Script / Orthography
eng_Latn
314-0.3
314.0
0.3
{ "CohereLabs/aya-expanse-8b": 0.9999998807907104, "Qwen/Qwen3-8B": 1, "bigscience/bloom": 0.9999998807907104, "common-pile/comma-v0.1-1t": 0.9999998807907104, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000001192092896, "google/byt5-small": 1.0000001192092896, "google/gemma-2-2b": 0.9999998211860657, "gpt2": 1, "meta-llama/Llama-3.2-1B": 1, "microsoft/Phi-3-mini-4k-instruct": 1.0000001192092896, "mistralai/tekken": 1, "tiktoken/gpt-4o": 0.9999998807907104, "tokenmonster/englishcode-32000-consistent-v1": 0.9999997615814209 }
{ "CohereLabs/aya-expanse-8b": 0.9999999403953552, "Qwen/Qwen3-8B": 1.000000238418579, "bigscience/bloom": 1.0000003576278687, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.0000004768371582, "google-bert/bert-base-multilingual-cased": 1.000000238418579, "google/byt5-small": 1.000000238418579, "google/gemma-2-2b": 1.000000238418579, "gpt2": 1.0000003576278687, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000004768371582, "mistralai/tekken": 1.0000003576278687, "tiktoken/gpt-4o": 1.0000003576278687, "tokenmonster/englishcode-32000-consistent-v1": 1.0000005960464478 }
{ "CohereLabs/aya-expanse-8b": 10, "Qwen/Qwen3-8B": 10, "bigscience/bloom": 10, "common-pile/comma-v0.1-1t": 10, "facebook/xglm-564M": 10, "google-bert/bert-base-multilingual-cased": 11, "google/byt5-small": 32, "google/gemma-2-2b": 10, "gpt2": 10, "meta-llama/Llama-3.2-1B": 10, "microsoft/Phi-3-mini-4k-instruct": 10, "mistralai/tekken": 10, "tiktoken/gpt-4o": 10, "tokenmonster/englishcode-32000-consistent-v1": 8 }
In "I work at Google", Google is a
[ "co.", "pers.", "cty.", "fr." ]
0
A
test
Abbreviations
Script / Orthography
eng_Latn
315-0.4
315.0
0.4
{ "CohereLabs/aya-expanse-8b": 1, "Qwen/Qwen3-8B": 1, "bigscience/bloom": 0.9999998807907104, "common-pile/comma-v0.1-1t": 1, "facebook/xglm-564M": 1, "google-bert/bert-base-multilingual-cased": 1.0000001192092896, "google/byt5-small": 1.0000001192092896, "google/gemma-2-2b": 0.9999998807907104, "gpt2": 1.0000001192092896, "meta-llama/Llama-3.2-1B": 0.9999999403953552, "microsoft/Phi-3-mini-4k-instruct": 1.0000001192092896, "mistralai/tekken": 0.9999998807907104, "tiktoken/gpt-4o": 1.0000001192092896, "tokenmonster/englishcode-32000-consistent-v1": 0.9999997615814209 }
{ "CohereLabs/aya-expanse-8b": 0.9999999403953552, "Qwen/Qwen3-8B": 1.000000238418579, "bigscience/bloom": 1.0000003576278687, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.0000004768371582, "google-bert/bert-base-multilingual-cased": 1.000000238418579, "google/byt5-small": 1.000000238418579, "google/gemma-2-2b": 1.000000238418579, "gpt2": 1.0000003576278687, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000004768371582, "mistralai/tekken": 1.0000003576278687, "tiktoken/gpt-4o": 1.0000003576278687, "tokenmonster/englishcode-32000-consistent-v1": 1.0000005960464478 }
{ "CohereLabs/aya-expanse-8b": 10, "Qwen/Qwen3-8B": 10, "bigscience/bloom": 10, "common-pile/comma-v0.1-1t": 10, "facebook/xglm-564M": 10, "google-bert/bert-base-multilingual-cased": 11, "google/byt5-small": 34, "google/gemma-2-2b": 10, "gpt2": 10, "meta-llama/Llama-3.2-1B": 10, "microsoft/Phi-3-mini-4k-instruct": 10, "mistralai/tekken": 10, "tiktoken/gpt-4o": 10, "tokenmonster/englishcode-32000-consistent-v1": 8 }
In "Microsoft released a new update", Microsoft is a
[ "pers.", "cty.", "co.", "fr." ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
316-0.5
316.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.9999996423721313, "Qwen/Qwen3-8B": 0.9999999403953552, "bigscience/bloom": 0.9999998807907104, "common-pile/comma-v0.1-1t": 1.0000001192092896, "facebook/xglm-564M": 0.9999998211860657, "google-bert/bert-base-multilingual-cased": 0.9999999403953552, "google/byt5-small": 1, "google/gemma-2-2b": 0.9999996423721313, "gpt2": 1.000000238418579, "meta-llama/Llama-3.2-1B": 0.9999999403953552, "microsoft/Phi-3-mini-4k-instruct": 1.0000001192092896, "mistralai/tekken": 1.0000001192092896, "tiktoken/gpt-4o": 1.0000001192092896, "tokenmonster/englishcode-32000-consistent-v1": 1.0000001192092896 }
{ "CohereLabs/aya-expanse-8b": 0.9999999403953552, "Qwen/Qwen3-8B": 1.000000238418579, "bigscience/bloom": 1.0000003576278687, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.0000004768371582, "google-bert/bert-base-multilingual-cased": 1.000000238418579, "google/byt5-small": 1.000000238418579, "google/gemma-2-2b": 1.000000238418579, "gpt2": 1.0000003576278687, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000004768371582, "mistralai/tekken": 1.0000003576278687, "tiktoken/gpt-4o": 1.0000003576278687, "tokenmonster/englishcode-32000-consistent-v1": 1.0000005960464478 }
{ "CohereLabs/aya-expanse-8b": 11, "Qwen/Qwen3-8B": 11, "bigscience/bloom": 12, "common-pile/comma-v0.1-1t": 11, "facebook/xglm-564M": 11, "google-bert/bert-base-multilingual-cased": 12, "google/byt5-small": 52, "google/gemma-2-2b": 11, "gpt2": 11, "meta-llama/Llama-3.2-1B": 11, "microsoft/Phi-3-mini-4k-instruct": 11, "mistralai/tekken": 11, "tiktoken/gpt-4o": 11, "tokenmonster/englishcode-32000-consistent-v1": 9 }
In "The cat sat on the mat", the subj. is
[ "the cat", "sat", "the mat", "on" ]
0
A
test
Abbreviations
Script / Orthography
eng_Latn
317-0.5
317.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.9349287152290344, "Qwen/Qwen3-8B": 0.9383683204650879, "bigscience/bloom": 0.9273239374160767, "common-pile/comma-v0.1-1t": 0.8940730094909668, "facebook/xglm-564M": 0.8844199776649475, "google-bert/bert-base-multilingual-cased": 0.8985786437988281, "google/byt5-small": 0.9816980957984924, "google/gemma-2-2b": 0.9320594668388367, "gpt2": 0.8804003000259399, "meta-llama/Llama-3.2-1B": 0.92516028881073, "microsoft/Phi-3-mini-4k-instruct": 0.8588052988052368, "mistralai/tekken": 0.8823502659797668, "tiktoken/gpt-4o": 0.9314396381378174, "tokenmonster/englishcode-32000-consistent-v1": 0.8803983330726624 }
{ "CohereLabs/aya-expanse-8b": 0.0772608071565628, "Qwen/Qwen3-8B": 0.10374213010072708, "bigscience/bloom": 0.08070938289165497, "common-pile/comma-v0.1-1t": 0.07346736639738083, "facebook/xglm-564M": 0.05824323743581772, "google-bert/bert-base-multilingual-cased": 0.08166754245758057, "google/byt5-small": -0.046926841139793396, "google/gemma-2-2b": 0.0717279389500618, "gpt2": 0.026471275836229324, "meta-llama/Llama-3.2-1B": 0.06531772017478943, "microsoft/Phi-3-mini-4k-instruct": -0.03549063205718994, "mistralai/tekken": 0.07283864170312881, "tiktoken/gpt-4o": 0.07994227856397629, "tokenmonster/englishcode-32000-consistent-v1": 0.05385168641805649 }
{ "CohereLabs/aya-expanse-8b": 13, "Qwen/Qwen3-8B": 13, "bigscience/bloom": 13, "common-pile/comma-v0.1-1t": 14, "facebook/xglm-564M": 14, "google-bert/bert-base-multilingual-cased": 15, "google/byt5-small": 41, "google/gemma-2-2b": 13, "gpt2": 14, "meta-llama/Llama-3.2-1B": 13, "microsoft/Phi-3-mini-4k-instruct": 14, "mistralai/tekken": 14, "tiktoken/gpt-4o": 13, "tokenmonster/englishcode-32000-consistent-v1": 12 }
The gas humans need to breathe to live is
[ "CH₄", "O₂", "He", "H₂" ]
1
B
test
Abbreviations
Script / Orthography
eng_Latn
322-0.5
322.0
0.5
{ "CohereLabs/aya-expanse-8b": 1.000000238418579, "Qwen/Qwen3-8B": 1.0000001192092896, "bigscience/bloom": 1.0000001192092896, "common-pile/comma-v0.1-1t": 1, "facebook/xglm-564M": 1.0000001192092896, "google-bert/bert-base-multilingual-cased": 0.9999999403953552, "google/byt5-small": 0.9999997615814209, "google/gemma-2-2b": 1.0000001192092896, "gpt2": 1.0000001192092896, "meta-llama/Llama-3.2-1B": 1.0000001192092896, "microsoft/Phi-3-mini-4k-instruct": 1.0000001192092896, "mistralai/tekken": 1.000000238418579, "tiktoken/gpt-4o": 1.000000238418579, "tokenmonster/englishcode-32000-consistent-v1": 0.9999999403953552 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687 }
{ "CohereLabs/aya-expanse-8b": 9, "Qwen/Qwen3-8B": 9, "bigscience/bloom": 9, "common-pile/comma-v0.1-1t": 14, "facebook/xglm-564M": 9, "google-bert/bert-base-multilingual-cased": 10, "google/byt5-small": 41, "google/gemma-2-2b": 9, "gpt2": 9, "meta-llama/Llama-3.2-1B": 9, "microsoft/Phi-3-mini-4k-instruct": 11, "mistralai/tekken": 9, "tiktoken/gpt-4o": 9, "tokenmonster/englishcode-32000-consistent-v1": 8 }
Chad's cap. is
[ "Moundou", "Abéché", "Ngama", "N'Djamena" ]
3
D
test
Abbreviations
Script / Orthography
eng_Latn
326-0.3
326.0
0.3
{ "CohereLabs/aya-expanse-8b": 0.696435809135437, "Qwen/Qwen3-8B": 0.7792327404022217, "bigscience/bloom": 0.7729296684265137, "common-pile/comma-v0.1-1t": 0.7977029085159302, "facebook/xglm-564M": 0.7329659461975098, "google-bert/bert-base-multilingual-cased": 0.7466813325881958, "google/byt5-small": 0.9061909317970276, "google/gemma-2-2b": 0.7286580204963684, "gpt2": 0.783542275428772, "meta-llama/Llama-3.2-1B": 0.7782202959060669, "microsoft/Phi-3-mini-4k-instruct": 0.8116836547851562, "mistralai/tekken": 0.7620350122451782, "tiktoken/gpt-4o": 0.781197190284729, "tokenmonster/englishcode-32000-consistent-v1": 0.8143059015274048 }
{ "CohereLabs/aya-expanse-8b": 0.1356840431690216, "Qwen/Qwen3-8B": 0.054030366241931915, "bigscience/bloom": 0.02676394209265709, "common-pile/comma-v0.1-1t": 0.05686108395457268, "facebook/xglm-564M": 0.06816975772380829, "google-bert/bert-base-multilingual-cased": 0.057269658893346786, "google/byt5-small": -0.02295071817934513, "google/gemma-2-2b": 0.09922298789024353, "gpt2": 0.10866591334342957, "meta-llama/Llama-3.2-1B": 0.09607133269309998, "microsoft/Phi-3-mini-4k-instruct": 0.13172483444213867, "mistralai/tekken": 0.04374687373638153, "tiktoken/gpt-4o": 0.08728465437889099, "tokenmonster/englishcode-32000-consistent-v1": 0.06787453591823578 }
{ "CohereLabs/aya-expanse-8b": 5, "Qwen/Qwen3-8B": 6, "bigscience/bloom": 6, "common-pile/comma-v0.1-1t": 6, "facebook/xglm-564M": 6, "google-bert/bert-base-multilingual-cased": 6, "google/byt5-small": 14, "google/gemma-2-2b": 6, "gpt2": 6, "meta-llama/Llama-3.2-1B": 6, "microsoft/Phi-3-mini-4k-instruct": 7, "mistralai/tekken": 6, "tiktoken/gpt-4o": 6, "tokenmonster/englishcode-32000-consistent-v1": 7 }
The capital of FR is
[ "London", "Berlin", "Rome", "Paris" ]
3
D
test
Abbreviations
Script / Orthography
eng_Latn
327-0.4
327.0
0.4
{ "CohereLabs/aya-expanse-8b": 0.7995725274085999, "Qwen/Qwen3-8B": 0.8451246023178101, "bigscience/bloom": 0.8283455967903137, "common-pile/comma-v0.1-1t": 0.8237259984016418, "facebook/xglm-564M": 0.806121826171875, "google-bert/bert-base-multilingual-cased": 0.7981747388839722, "google/byt5-small": 0.9172342419624329, "google/gemma-2-2b": 0.7995125651359558, "gpt2": 0.8134028315544128, "meta-llama/Llama-3.2-1B": 0.8076421618461609, "microsoft/Phi-3-mini-4k-instruct": 0.8156378865242004, "mistralai/tekken": 0.7921538352966309, "tiktoken/gpt-4o": 0.7901878356933594, "tokenmonster/englishcode-32000-consistent-v1": 0.7765400409698486 }
{ "CohereLabs/aya-expanse-8b": 0.08211846649646759, "Qwen/Qwen3-8B": 0.14105814695358276, "bigscience/bloom": 0.1934700310230255, "common-pile/comma-v0.1-1t": 0.09907892346382141, "facebook/xglm-564M": 0.08311900496482849, "google-bert/bert-base-multilingual-cased": 0.13695181906223297, "google/byt5-small": 0.14798158407211304, "google/gemma-2-2b": 0.01552680041640997, "gpt2": 0.14244207739830017, "meta-llama/Llama-3.2-1B": 0.14876556396484375, "microsoft/Phi-3-mini-4k-instruct": 0.1234123557806015, "mistralai/tekken": 0.08589836210012436, "tiktoken/gpt-4o": 0.07952668517827988, "tokenmonster/englishcode-32000-consistent-v1": 0.026012370362877846 }
{ "CohereLabs/aya-expanse-8b": 5, "Qwen/Qwen3-8B": 5, "bigscience/bloom": 5, "common-pile/comma-v0.1-1t": 6, "facebook/xglm-564M": 5, "google-bert/bert-base-multilingual-cased": 5, "google/byt5-small": 20, "google/gemma-2-2b": 5, "gpt2": 5, "meta-llama/Llama-3.2-1B": 5, "microsoft/Phi-3-mini-4k-instruct": 5, "mistralai/tekken": 5, "tiktoken/gpt-4o": 5, "tokenmonster/englishcode-32000-consistent-v1": 6 }
The capital of JP is
[ "Kyoto", "Tokyo", "Osaka", "Hiroshima" ]
1
B
test
Abbreviations
Script / Orthography
eng_Latn
328-0.4
328.0
0.4
{ "CohereLabs/aya-expanse-8b": 0.831386148929596, "Qwen/Qwen3-8B": 0.8522142171859741, "bigscience/bloom": 0.8232855200767517, "common-pile/comma-v0.1-1t": 0.8368542194366455, "facebook/xglm-564M": 0.8285537362098694, "google-bert/bert-base-multilingual-cased": 0.8010603785514832, "google/byt5-small": 0.9203615784645081, "google/gemma-2-2b": 0.8251798152923584, "gpt2": 0.8022357225418091, "meta-llama/Llama-3.2-1B": 0.8143433332443237, "microsoft/Phi-3-mini-4k-instruct": 0.7175941467285156, "mistralai/tekken": 0.8012662529945374, "tiktoken/gpt-4o": 0.8201245069503784, "tokenmonster/englishcode-32000-consistent-v1": 0.8073024749755859 }
{ "CohereLabs/aya-expanse-8b": 0.1724080741405487, "Qwen/Qwen3-8B": 0.12307695299386978, "bigscience/bloom": 0.08853715658187866, "common-pile/comma-v0.1-1t": 0.11716070026159286, "facebook/xglm-564M": 0.15991993248462677, "google-bert/bert-base-multilingual-cased": 0.08823803812265396, "google/byt5-small": 0.2070769965648651, "google/gemma-2-2b": 0.06976078450679779, "gpt2": 0.08312174677848816, "meta-llama/Llama-3.2-1B": 0.12662671506404877, "microsoft/Phi-3-mini-4k-instruct": 0.13243405520915985, "mistralai/tekken": 0.06049402803182602, "tiktoken/gpt-4o": 0.14071625471115112, "tokenmonster/englishcode-32000-consistent-v1": 0.10549987852573395 }
{ "CohereLabs/aya-expanse-8b": 5, "Qwen/Qwen3-8B": 5, "bigscience/bloom": 5, "common-pile/comma-v0.1-1t": 6, "facebook/xglm-564M": 5, "google-bert/bert-base-multilingual-cased": 5, "google/byt5-small": 20, "google/gemma-2-2b": 5, "gpt2": 5, "meta-llama/Llama-3.2-1B": 5, "microsoft/Phi-3-mini-4k-instruct": 6, "mistralai/tekken": 5, "tiktoken/gpt-4o": 5, "tokenmonster/englishcode-32000-consistent-v1": 6 }
The capital of TR is
[ "İstanbul", "İzmir", "Ankara", "Bursa" ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
329-0.3
329.0
0.3
{ "CohereLabs/aya-expanse-8b": 0.8062456846237183, "Qwen/Qwen3-8B": 0.8371235132217407, "bigscience/bloom": 0.8147814869880676, "common-pile/comma-v0.1-1t": 0.8117464184761047, "facebook/xglm-564M": 0.8058038949966431, "google-bert/bert-base-multilingual-cased": 0.807705283164978, "google/byt5-small": 0.8975408673286438, "google/gemma-2-2b": 0.8226367831230164, "gpt2": 0.7932828664779663, "meta-llama/Llama-3.2-1B": 0.8048302531242371, "microsoft/Phi-3-mini-4k-instruct": 0.7885987758636475, "mistralai/tekken": 0.8118667602539062, "tiktoken/gpt-4o": 0.7969956994056702, "tokenmonster/englishcode-32000-consistent-v1": 0.7851678729057312 }
{ "CohereLabs/aya-expanse-8b": 0.08540067076683044, "Qwen/Qwen3-8B": 0.09242145717144012, "bigscience/bloom": 0.11954235285520554, "common-pile/comma-v0.1-1t": 0.0688096210360527, "facebook/xglm-564M": 0.056273337453603745, "google-bert/bert-base-multilingual-cased": 0.184139221906662, "google/byt5-small": 0.12901875376701355, "google/gemma-2-2b": 0.11601769179105759, "gpt2": 0.07733984291553497, "meta-llama/Llama-3.2-1B": 0.14307504892349243, "microsoft/Phi-3-mini-4k-instruct": -0.00042920373380184174, "mistralai/tekken": 0.1253935694694519, "tiktoken/gpt-4o": 0.04226153716444969, "tokenmonster/englishcode-32000-consistent-v1": -0.0061051626689732075 }
{ "CohereLabs/aya-expanse-8b": 5, "Qwen/Qwen3-8B": 5, "bigscience/bloom": 5, "common-pile/comma-v0.1-1t": 6, "facebook/xglm-564M": 5, "google-bert/bert-base-multilingual-cased": 5, "google/byt5-small": 20, "google/gemma-2-2b": 5, "gpt2": 5, "meta-llama/Llama-3.2-1B": 5, "microsoft/Phi-3-mini-4k-instruct": 5, "mistralai/tekken": 5, "tiktoken/gpt-4o": 5, "tokenmonster/englishcode-32000-consistent-v1": 6 }
The chem. formula for water is
[ "H2O", "CO2", "NaCl", "O2" ]
0
A
test
Abbreviations
Script / Orthography
eng_Latn
330-0.4
330.0
0.4
{ "CohereLabs/aya-expanse-8b": 0.8186859488487244, "Qwen/Qwen3-8B": 0.8713029026985168, "bigscience/bloom": 0.8499932885169983, "common-pile/comma-v0.1-1t": 0.8503109216690063, "facebook/xglm-564M": 0.7378727793693542, "google-bert/bert-base-multilingual-cased": 0.7587925791740417, "google/byt5-small": 0.9591166973114014, "google/gemma-2-2b": 0.8184010982513428, "gpt2": 0.8338674306869507, "meta-llama/Llama-3.2-1B": 0.819198727607727, "microsoft/Phi-3-mini-4k-instruct": 0.8195061683654785, "mistralai/tekken": 0.8313064575195312, "tiktoken/gpt-4o": 0.836508572101593, "tokenmonster/englishcode-32000-consistent-v1": 0.848869264125824 }
{ "CohereLabs/aya-expanse-8b": 0.19135701656341553, "Qwen/Qwen3-8B": 0.21313846111297607, "bigscience/bloom": 0.23347987234592438, "common-pile/comma-v0.1-1t": 0.18952646851539612, "facebook/xglm-564M": 0.021598808467388153, "google-bert/bert-base-multilingual-cased": 0.016012798994779587, "google/byt5-small": -0.02367701567709446, "google/gemma-2-2b": 0.17226368188858032, "gpt2": 0.20595847070217133, "meta-llama/Llama-3.2-1B": 0.2124955654144287, "microsoft/Phi-3-mini-4k-instruct": 0.15150737762451172, "mistralai/tekken": 0.24085702002048492, "tiktoken/gpt-4o": 0.16686344146728516, "tokenmonster/englishcode-32000-consistent-v1": 0.14736664295196533 }
{ "CohereLabs/aya-expanse-8b": 7, "Qwen/Qwen3-8B": 7, "bigscience/bloom": 7, "common-pile/comma-v0.1-1t": 7, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 30, "google/gemma-2-2b": 7, "gpt2": 7, "meta-llama/Llama-3.2-1B": 7, "microsoft/Phi-3-mini-4k-instruct": 7, "mistralai/tekken": 7, "tiktoken/gpt-4o": 7, "tokenmonster/englishcode-32000-consistent-v1": 7 }
The intent in "What time does the store close?" is
[ "purch", "book", "complain", "info" ]
3
D
test
Abbreviations
Script / Orthography
eng_Latn
331-0.4
331.0
0.4
{ "CohereLabs/aya-expanse-8b": 0.9999998807907104, "Qwen/Qwen3-8B": 1.0000001192092896, "bigscience/bloom": 0.9999998807907104, "common-pile/comma-v0.1-1t": 1.0000001192092896, "facebook/xglm-564M": 1, "google-bert/bert-base-multilingual-cased": 1.0000001192092896, "google/byt5-small": 1, "google/gemma-2-2b": 1, "gpt2": 1, "meta-llama/Llama-3.2-1B": 0.9999995827674866, "microsoft/Phi-3-mini-4k-instruct": 0.9999997615814209, "mistralai/tekken": 1, "tiktoken/gpt-4o": 0.9999998807907104, "tokenmonster/englishcode-32000-consistent-v1": 0.9999999403953552 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687 }
{ "CohereLabs/aya-expanse-8b": 12, "Qwen/Qwen3-8B": 12, "bigscience/bloom": 12, "common-pile/comma-v0.1-1t": 13, "facebook/xglm-564M": 12, "google-bert/bert-base-multilingual-cased": 13, "google/byt5-small": 50, "google/gemma-2-2b": 12, "gpt2": 12, "meta-llama/Llama-3.2-1B": 12, "microsoft/Phi-3-mini-4k-instruct": 12, "mistralai/tekken": 12, "tiktoken/gpt-4o": 12, "tokenmonster/englishcode-32000-consistent-v1": 11 }
The largest mammal in the world is
[ "dolphin", "giraffe", "blue whale", "bear" ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
332-0.5
332.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.9999997615814209, "Qwen/Qwen3-8B": 0.9999997615814209, "bigscience/bloom": 0.9999997019767761, "common-pile/comma-v0.1-1t": 1, "facebook/xglm-564M": 0.9999997615814209, "google-bert/bert-base-multilingual-cased": 1, "google/byt5-small": 0.9999998211860657, "google/gemma-2-2b": 0.9999999403953552, "gpt2": 1, "meta-llama/Llama-3.2-1B": 0.9999998211860657, "microsoft/Phi-3-mini-4k-instruct": 0.9999998807907104, "mistralai/tekken": 1, "tiktoken/gpt-4o": 0.9999997615814209, "tokenmonster/englishcode-32000-consistent-v1": 0.9999998807907104 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687 }
{ "CohereLabs/aya-expanse-8b": 7, "Qwen/Qwen3-8B": 8, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 8, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 34, "google/gemma-2-2b": 7, "gpt2": 7, "meta-llama/Llama-3.2-1B": 8, "microsoft/Phi-3-mini-4k-instruct": 9, "mistralai/tekken": 8, "tiktoken/gpt-4o": 8, "tokenmonster/englishcode-32000-consistent-v1": 7 }
The unit of measurement for temperature in the International System is
[ "°C", "m", "K", "°R" ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
333-0.4
333.0
0.4
{ "CohereLabs/aya-expanse-8b": 1, "Qwen/Qwen3-8B": 1.0000001192092896, "bigscience/bloom": 0.9999998807907104, "common-pile/comma-v0.1-1t": 1, "facebook/xglm-564M": 0.9999998807907104, "google-bert/bert-base-multilingual-cased": 0.9999998211860657, "google/byt5-small": 1, "google/gemma-2-2b": 0.9999999403953552, "gpt2": 1.0000001192092896, "meta-llama/Llama-3.2-1B": 1, "microsoft/Phi-3-mini-4k-instruct": 1, "mistralai/tekken": 1.0000001192092896, "tiktoken/gpt-4o": 1.0000001192092896, "tokenmonster/englishcode-32000-consistent-v1": 0.9999999403953552 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.000000238418579 }
{ "CohereLabs/aya-expanse-8b": 11, "Qwen/Qwen3-8B": 11, "bigscience/bloom": 11, "common-pile/comma-v0.1-1t": 13, "facebook/xglm-564M": 11, "google-bert/bert-base-multilingual-cased": 11, "google/byt5-small": 70, "google/gemma-2-2b": 11, "gpt2": 11, "meta-llama/Llama-3.2-1B": 11, "microsoft/Phi-3-mini-4k-instruct": 11, "mistralai/tekken": 11, "tiktoken/gpt-4o": 11, "tokenmonster/englishcode-32000-consistent-v1": 10 }
The country whose space agency is NASA is
[ "RU", "US", "CN", "JP" ]
1
B
test
Abbreviations
Script / Orthography
eng_Latn
334-0.3
334.0
0.3
{ "CohereLabs/aya-expanse-8b": 0.9999997615814209, "Qwen/Qwen3-8B": 0.9999998807907104, "bigscience/bloom": 0.9999997019767761, "common-pile/comma-v0.1-1t": 0.9999997615814209, "facebook/xglm-564M": 0.9999997615814209, "google-bert/bert-base-multilingual-cased": 1, "google/byt5-small": 0.9999998211860657, "google/gemma-2-2b": 0.9999997615814209, "gpt2": 1.0000001192092896, "meta-llama/Llama-3.2-1B": 0.9999998807907104, "microsoft/Phi-3-mini-4k-instruct": 1.000000238418579, "mistralai/tekken": 0.9999996423721313, "tiktoken/gpt-4o": 0.9999998807907104, "tokenmonster/englishcode-32000-consistent-v1": 1.0000001192092896 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687 }
{ "CohereLabs/aya-expanse-8b": 8, "Qwen/Qwen3-8B": 8, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 8, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 41, "google/gemma-2-2b": 8, "gpt2": 8, "meta-llama/Llama-3.2-1B": 8, "microsoft/Phi-3-mini-4k-instruct": 9, "mistralai/tekken": 8, "tiktoken/gpt-4o": 8, "tokenmonster/englishcode-32000-consistent-v1": 9 }
The lang. spoken in Brazil is
[ "Spanish", "French", "Italian", "Portuguese" ]
3
D
test
Abbreviations
Script / Orthography
eng_Latn
335-0.5
335.0
0.5
{ "CohereLabs/aya-expanse-8b": 0.7913755178451538, "Qwen/Qwen3-8B": 0.8113056421279907, "bigscience/bloom": 0.8049988746643066, "common-pile/comma-v0.1-1t": 0.8325293064117432, "facebook/xglm-564M": 0.7717224359512329, "google-bert/bert-base-multilingual-cased": 0.7581318616867065, "google/byt5-small": 0.9475886821746826, "google/gemma-2-2b": 0.7959599494934082, "gpt2": 0.788682222366333, "meta-llama/Llama-3.2-1B": 0.8063963055610657, "microsoft/Phi-3-mini-4k-instruct": 0.7926681041717529, "mistralai/tekken": 0.7794783711433411, "tiktoken/gpt-4o": 0.8209499716758728, "tokenmonster/englishcode-32000-consistent-v1": 0.7077274322509766 }
{ "CohereLabs/aya-expanse-8b": 0.10002417862415314, "Qwen/Qwen3-8B": 0.07510721683502197, "bigscience/bloom": 0.12772142887115479, "common-pile/comma-v0.1-1t": 0.06116257607936859, "facebook/xglm-564M": 0.09400343149900436, "google-bert/bert-base-multilingual-cased": 0.024091584607958794, "google/byt5-small": -0.025367457419633865, "google/gemma-2-2b": 0.14458352327346802, "gpt2": 0.10195167362689972, "meta-llama/Llama-3.2-1B": 0.11951098591089249, "microsoft/Phi-3-mini-4k-instruct": 0.04442334175109863, "mistralai/tekken": 0.07316195219755173, "tiktoken/gpt-4o": 0.16807381808757782, "tokenmonster/englishcode-32000-consistent-v1": 0.16010284423828125 }
{ "CohereLabs/aya-expanse-8b": 7, "Qwen/Qwen3-8B": 7, "bigscience/bloom": 7, "common-pile/comma-v0.1-1t": 8, "facebook/xglm-564M": 7, "google-bert/bert-base-multilingual-cased": 7, "google/byt5-small": 29, "google/gemma-2-2b": 7, "gpt2": 7, "meta-llama/Llama-3.2-1B": 7, "microsoft/Phi-3-mini-4k-instruct": 7, "mistralai/tekken": 7, "tiktoken/gpt-4o": 7, "tokenmonster/englishcode-32000-consistent-v1": 7 }
The metal with chemical sym. 'Fe' is
[ "lead", "zinc", "iron", "gold" ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
336-0.3
336.0
0.3
{ "CohereLabs/aya-expanse-8b": 0.9029386043548584, "Qwen/Qwen3-8B": 0.9102910757064819, "bigscience/bloom": 0.8799982070922852, "common-pile/comma-v0.1-1t": 0.9248884916305542, "facebook/xglm-564M": 0.8538807034492493, "google-bert/bert-base-multilingual-cased": 0.8467434644699097, "google/byt5-small": 0.9709159731864929, "google/gemma-2-2b": 0.9017316699028015, "gpt2": 0.8987561464309692, "meta-llama/Llama-3.2-1B": 0.9013137221336365, "microsoft/Phi-3-mini-4k-instruct": 0.8870991468429565, "mistralai/tekken": 0.8897440433502197, "tiktoken/gpt-4o": 0.913067102432251, "tokenmonster/englishcode-32000-consistent-v1": 0.868256688117981 }
{ "CohereLabs/aya-expanse-8b": 0.14960479736328125, "Qwen/Qwen3-8B": 0.10851742327213287, "bigscience/bloom": 0.05709414184093475, "common-pile/comma-v0.1-1t": 0.03906556963920593, "facebook/xglm-564M": 0.06293395161628723, "google-bert/bert-base-multilingual-cased": 0.03629113733768463, "google/byt5-small": -0.034767430275678635, "google/gemma-2-2b": 0.10764612257480621, "gpt2": 0.08282459527254105, "meta-llama/Llama-3.2-1B": 0.1575041115283966, "microsoft/Phi-3-mini-4k-instruct": 0.050431989133358, "mistralai/tekken": 0.10146505385637283, "tiktoken/gpt-4o": 0.2269553542137146, "tokenmonster/englishcode-32000-consistent-v1": -0.060447804629802704 }
{ "CohereLabs/aya-expanse-8b": 10, "Qwen/Qwen3-8B": 10, "bigscience/bloom": 10, "common-pile/comma-v0.1-1t": 15, "facebook/xglm-564M": 11, "google-bert/bert-base-multilingual-cased": 11, "google/byt5-small": 36, "google/gemma-2-2b": 10, "gpt2": 10, "meta-llama/Llama-3.2-1B": 10, "microsoft/Phi-3-mini-4k-instruct": 10, "mistralai/tekken": 10, "tiktoken/gpt-4o": 10, "tokenmonster/englishcode-32000-consistent-v1": 10 }
The planet closest to the Sun in our solar system is
[ "♀", "♂", "☿", "♁" ]
2
C
test
Abbreviations
Script / Orthography
eng_Latn
338-0.4
338.0
0.4
{ "CohereLabs/aya-expanse-8b": 0.9999999403953552, "Qwen/Qwen3-8B": 1, "bigscience/bloom": 0.9999999403953552, "common-pile/comma-v0.1-1t": 1.0000001192092896, "facebook/xglm-564M": 0.9999999403953552, "google-bert/bert-base-multilingual-cased": 0.9999997019767761, "google/byt5-small": 1, "google/gemma-2-2b": 0.9999999403953552, "gpt2": 1, "meta-llama/Llama-3.2-1B": 0.9999998807907104, "microsoft/Phi-3-mini-4k-instruct": 0.9999999403953552, "mistralai/tekken": 0.9999999403953552, "tiktoken/gpt-4o": 1.0000001192092896, "tokenmonster/englishcode-32000-consistent-v1": 0.9999998211860657 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.000000238418579 }
{ "CohereLabs/aya-expanse-8b": 11, "Qwen/Qwen3-8B": 11, "bigscience/bloom": 11, "common-pile/comma-v0.1-1t": 14, "facebook/xglm-564M": 12, "google-bert/bert-base-multilingual-cased": 11, "google/byt5-small": 52, "google/gemma-2-2b": 11, "gpt2": 11, "meta-llama/Llama-3.2-1B": 11, "microsoft/Phi-3-mini-4k-instruct": 11, "mistralai/tekken": 11, "tiktoken/gpt-4o": 11, "tokenmonster/englishcode-32000-consistent-v1": 8 }
The largest planet in the Solar System is
[ "♁", "♃", "♄", "♂" ]
1
B
test
Abbreviations
Script / Orthography
eng_Latn
339-0.4
339.0
0.4
{ "CohereLabs/aya-expanse-8b": 0.9999998211860657, "Qwen/Qwen3-8B": 0.9999998211860657, "bigscience/bloom": 0.9999997615814209, "common-pile/comma-v0.1-1t": 1.0000001192092896, "facebook/xglm-564M": 1, "google-bert/bert-base-multilingual-cased": 0.9999999403953552, "google/byt5-small": 1.000000238418579, "google/gemma-2-2b": 1, "gpt2": 0.9999996423721313, "meta-llama/Llama-3.2-1B": 1, "microsoft/Phi-3-mini-4k-instruct": 1, "mistralai/tekken": 0.9999998211860657, "tiktoken/gpt-4o": 0.9999998211860657, "tokenmonster/englishcode-32000-consistent-v1": 0.9999998211860657 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.000000238418579 }
{ "CohereLabs/aya-expanse-8b": 8, "Qwen/Qwen3-8B": 8, "bigscience/bloom": 8, "common-pile/comma-v0.1-1t": 9, "facebook/xglm-564M": 8, "google-bert/bert-base-multilingual-cased": 8, "google/byt5-small": 41, "google/gemma-2-2b": 8, "gpt2": 8, "meta-llama/Llama-3.2-1B": 8, "microsoft/Phi-3-mini-4k-instruct": 9, "mistralai/tekken": 8, "tiktoken/gpt-4o": 8, "tokenmonster/englishcode-32000-consistent-v1": 8 }
The process that allows plants to prod. their own food using sunlight is
[ "photosynthesis", "respiration", "digestion", "fermentation" ]
0
A
test
Abbreviations
Script / Orthography
eng_Latn
340-0.4
340.0
0.4
{ "CohereLabs/aya-expanse-8b": 0.9140673875808716, "Qwen/Qwen3-8B": 0.9345059990882874, "bigscience/bloom": 0.9172125458717346, "common-pile/comma-v0.1-1t": 0.9256795048713684, "facebook/xglm-564M": 0.9292510747909546, "google-bert/bert-base-multilingual-cased": 0.8929300308227539, "google/byt5-small": 0.993126630783081, "google/gemma-2-2b": 0.9281373620033264, "gpt2": 0.9144498705863953, "meta-llama/Llama-3.2-1B": 0.9152953624725342, "microsoft/Phi-3-mini-4k-instruct": 0.9282230138778687, "mistralai/tekken": 0.917586088180542, "tiktoken/gpt-4o": 0.9270635843276978, "tokenmonster/englishcode-32000-consistent-v1": 0.8559166193008423 }
{ "CohereLabs/aya-expanse-8b": 0.0885179415345192, "Qwen/Qwen3-8B": 0.19490374624729156, "bigscience/bloom": 0.0557279996573925, "common-pile/comma-v0.1-1t": 0.08006878197193146, "facebook/xglm-564M": 0.11164054274559021, "google-bert/bert-base-multilingual-cased": 0.0015388152096420527, "google/byt5-small": -0.029237380251288414, "google/gemma-2-2b": 0.14768964052200317, "gpt2": 0.08957405388355255, "meta-llama/Llama-3.2-1B": 0.0850566104054451, "microsoft/Phi-3-mini-4k-instruct": 0.06736037880182266, "mistralai/tekken": 0.07164368778467178, "tiktoken/gpt-4o": 0.11323671042919159, "tokenmonster/englishcode-32000-consistent-v1": -0.03460094332695007 }
{ "CohereLabs/aya-expanse-8b": 14, "Qwen/Qwen3-8B": 14, "bigscience/bloom": 14, "common-pile/comma-v0.1-1t": 16, "facebook/xglm-564M": 15, "google-bert/bert-base-multilingual-cased": 16, "google/byt5-small": 72, "google/gemma-2-2b": 14, "gpt2": 14, "meta-llama/Llama-3.2-1B": 14, "microsoft/Phi-3-mini-4k-instruct": 15, "mistralai/tekken": 14, "tiktoken/gpt-4o": 14, "tokenmonster/englishcode-32000-consistent-v1": 12 }
The author who wrote the play "Romeo and Juliet" is
[ "W. Shakespeare", "C. Dickens", "M. Twain", "J. Austen" ]
0
A
test
Abbreviations
Script / Orthography
eng_Latn
341-0.2
341.0
0.2
{ "CohereLabs/aya-expanse-8b": 0.9999999403953552, "Qwen/Qwen3-8B": 1.0000001192092896, "bigscience/bloom": 1.0000001192092896, "common-pile/comma-v0.1-1t": 1.0000001192092896, "facebook/xglm-564M": 1.0000001192092896, "google-bert/bert-base-multilingual-cased": 0.9999998211860657, "google/byt5-small": 1.0000001192092896, "google/gemma-2-2b": 1, "gpt2": 1, "meta-llama/Llama-3.2-1B": 0.9999999403953552, "microsoft/Phi-3-mini-4k-instruct": 1, "mistralai/tekken": 1, "tiktoken/gpt-4o": 1.000000238418579, "tokenmonster/englishcode-32000-consistent-v1": 1 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687 }
{ "CohereLabs/aya-expanse-8b": 12, "Qwen/Qwen3-8B": 14, "bigscience/bloom": 13, "common-pile/comma-v0.1-1t": 17, "facebook/xglm-564M": 14, "google-bert/bert-base-multilingual-cased": 12, "google/byt5-small": 51, "google/gemma-2-2b": 12, "gpt2": 14, "meta-llama/Llama-3.2-1B": 14, "microsoft/Phi-3-mini-4k-instruct": 15, "mistralai/tekken": 13, "tiktoken/gpt-4o": 13, "tokenmonster/englishcode-32000-consistent-v1": 14 }
What plants need from the air to make food is
[ "N₂", "H₂", "He", "CO₂" ]
3
D
test
Abbreviations
Script / Orthography
eng_Latn
343-0.3
343.0
0.3
{ "CohereLabs/aya-expanse-8b": 1, "Qwen/Qwen3-8B": 0.9999999403953552, "bigscience/bloom": 1.0000001192092896, "common-pile/comma-v0.1-1t": 0.9999998211860657, "facebook/xglm-564M": 1, "google-bert/bert-base-multilingual-cased": 0.9999998807907104, "google/byt5-small": 1, "google/gemma-2-2b": 0.9999997615814209, "gpt2": 1.0000001192092896, "meta-llama/Llama-3.2-1B": 1.0000001192092896, "microsoft/Phi-3-mini-4k-instruct": 0.9999998807907104, "mistralai/tekken": 1, "tiktoken/gpt-4o": 0.9999998211860657, "tokenmonster/englishcode-32000-consistent-v1": 0.9999998807907104 }
{ "CohereLabs/aya-expanse-8b": 1.0000003576278687, "Qwen/Qwen3-8B": 1.0000004768371582, "bigscience/bloom": 1.000000238418579, "common-pile/comma-v0.1-1t": 1.0000004768371582, "facebook/xglm-564M": 1.000000238418579, "google-bert/bert-base-multilingual-cased": 1.0000004768371582, "google/byt5-small": 1.0000004768371582, "google/gemma-2-2b": 1.0000003576278687, "gpt2": 1.0000004768371582, "meta-llama/Llama-3.2-1B": 1.0000003576278687, "microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687, "mistralai/tekken": 1.0000004768371582, "tiktoken/gpt-4o": 1.0000004768371582, "tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687 }
{ "CohereLabs/aya-expanse-8b": 10, "Qwen/Qwen3-8B": 10, "bigscience/bloom": 10, "common-pile/comma-v0.1-1t": 12, "facebook/xglm-564M": 10, "google-bert/bert-base-multilingual-cased": 10, "google/byt5-small": 45, "google/gemma-2-2b": 10, "gpt2": 10, "meta-llama/Llama-3.2-1B": 10, "microsoft/Phi-3-mini-4k-instruct": 10, "mistralai/tekken": 10, "tiktoken/gpt-4o": 10, "tokenmonster/englishcode-32000-consistent-v1": 8 }
In "Can you pls. book a flight to Paris?", the person wants to
[ "go shopping", "file a complaint", "cancel reservation", "make a booking" ]
3
D
test
Abbreviations
Script / Orthography
eng_Latn
344-0.4
344.0
0.4
{ "CohereLabs/aya-expanse-8b": 0.8967032432556152, "Qwen/Qwen3-8B": 0.9513365030288696, "bigscience/bloom": 0.9068660140037537, "common-pile/comma-v0.1-1t": 0.958304226398468, "facebook/xglm-564M": 0.9203377366065979, "google-bert/bert-base-multilingual-cased": 0.9147680997848511, "google/byt5-small": 0.9851800203323364, "google/gemma-2-2b": 0.9473385214805603, "gpt2": 0.8919826149940491, "meta-llama/Llama-3.2-1B": 0.9390807747840881, "microsoft/Phi-3-mini-4k-instruct": 0.9023863077163696, "mistralai/tekken": 0.884173572063446, "tiktoken/gpt-4o": 0.9475116729736328, "tokenmonster/englishcode-32000-consistent-v1": 0.8966159820556641 }
{ "CohereLabs/aya-expanse-8b": -0.009711533784866333, "Qwen/Qwen3-8B": 0.1350504606962204, "bigscience/bloom": 0.01425527036190033, "common-pile/comma-v0.1-1t": 0.13568083941936493, "facebook/xglm-564M": 0.06789200007915497, "google-bert/bert-base-multilingual-cased": 0.08616456389427185, "google/byt5-small": 0.30707496404647827, "google/gemma-2-2b": 0.1704685240983963, "gpt2": 0.0131788719445467, "meta-llama/Llama-3.2-1B": 0.140251025557518, "microsoft/Phi-3-mini-4k-instruct": 0.04106447473168373, "mistralai/tekken": 0.01378481462597847, "tiktoken/gpt-4o": 0.10346754640340805, "tokenmonster/englishcode-32000-consistent-v1": 0.07150956243276596 }
{ "CohereLabs/aya-expanse-8b": 17, "Qwen/Qwen3-8B": 16, "bigscience/bloom": 18, "common-pile/comma-v0.1-1t": 20, "facebook/xglm-564M": 18, "google-bert/bert-base-multilingual-cased": 18, "google/byt5-small": 62, "google/gemma-2-2b": 16, "gpt2": 17, "meta-llama/Llama-3.2-1B": 16, "microsoft/Phi-3-mini-4k-instruct": 18, "mistralai/tekken": 17, "tiktoken/gpt-4o": 16, "tokenmonster/englishcode-32000-consistent-v1": 14 }

Dataset Card for Tokenization Robustness

TokSuite Logo

TokSuite Benchmark (English Collection)

Dataset Description

This dataset is part of TokSuite, a comprehensive benchmark designed to measure how different tokenization strategies affect language model performance and robustness in isolation. This specific collection contains English multiple-choice text completion questions paired with a wide range of real-world surface-form perturbations that are known to interact strongly with tokenization.

  • Curated by: R3 Research Team
  • Language(s): English (en)
  • License: MIT License

Dataset Summary

TokSuite addresses a core challenge in language model research: isolating and measuring the impact of tokenizer choice on model behavior. The English collection serves as the reference and anchor language for TokSuite, providing a high-resource baseline with diverse perturbations that generalize across domains and writing styles.

Key Features:

  • 40 canonical English questions with high baseline accuracy
  • Extensive perturbation coverage spanning typography, formatting, morphology, noise, and stylistic variation
  • Parallel structure with TokSuite benchmarks in Turkish, Italian, Farsi, and Chinese
  • Controlled design enabling clean measurement of performance degradation under perturbations

Supported Tasks

  • Multiple-Choice Question Answering: Text completion with four answer options
  • Tokenizer Robustness Evaluation: Measuring accuracy drop under token-altering perturbations
  • Benchmarking Tokenization Effects: Isolating tokenizer behavior independent of model architecture or scale

Languages

The dataset contains text exclusively in English (language code: en).


Dataset Structure

Data Fields

Field Type Description
question string The question text in English
choices list[string] Four multiple-choice answer options
answer int64 Index of the correct answer
answer_label string Letter label of the correct answer
split string Dataset split identifier
subcategories string Perturbation category
lang string Language code (en)
second_lang string Optional paraphrase or descriptive reference
notes string Additional context about the perturbation
id string Unique question identifier
set_id float64 Question set grouping identifier
variation_id float64 Variation number within a question set
vanilla_cos_sim_to_canonical dict[string, float] Cosine similarity to canonical form (raw tokens)
trimmed_cos_sim_to_canonical dict[string, float] Cosine similarity after token normalization
token_counts dict[string, integer] Token counts per tokenizer

Dataset Creation

Curation Rationale

The English benchmark was created to:

  1. Serve as a high-resource reference language for tokenizer robustness studies
  2. Systematically probe tokenizer sensitivity to formatting, noise, and stylistic variation
  3. Enable controlled comparisons across tokenizers under identical model conditions
  4. Provide a reusable evaluation suite for studying tokenization effects in isolation

All canonical questions are intentionally simple, ensuring high baseline accuracy so that observed performance changes are attributable to perturbations rather than reasoning difficulty.

Source Data

Data Collection and Processing

  • Canonical Questions: 40 English questions authored by the TokSuite team
  • Perturbations: Targeted surface-form transformations applied per question
  • Validation: Model-in-the-loop verification to ensure canonical solvability

Perturbation Categories (English)

Each perturbation represents a distinct, realistic transformation of English text that can alter token boundaries or distributions.

  1. Canonical
    Standard, grammatically correct English text with no perturbations. Serves as the reference condition.

  2. Abbreviations
    Introduces common English abbreviations and shortened forms (e.g., Dr., etc., vs.).

  3. Capitalization
    Alters casing patterns through random capitalization, lowercasing, or mixed case.

  4. Character Deletion
    Removes characters within words, simulating typing omissions.

  5. Character Substitution
    Replaces characters with visually or keyboard-adjacent alternatives.

  6. Colloquial
    Applies informal spoken English forms and casual phrasing.

  7. Compounds
    Merges multi-word expressions into compound forms (e.g., notebook vs. note book).

  8. Contractions
    Uses contracted forms such as don’t, it’s, and they’re.

  9. Date Formats
    Varies date representations (e.g., March 12, 2022, 12/03/22, 2022-03-12).

  10. Emoji Substitution
    Replaces words with semantically related emojis.

  11. Grammatical Errors
    Injects plausible agreement, tense, or syntactic errors.

  12. Historical Spelling
    Uses archaic or historical English spellings.

  13. Homoglyphs
    Substitutes characters with visually similar Unicode glyphs.

  14. Hyphenated Spelling
    Introduces or removes hyphens in compound words.

  15. Inflections
    Alters tense, plurality, or derivational morphology.

  16. Keyboard Proximity Errors
    Simulates typos from adjacent keyboard keys.

  17. Letter Repetition for Emphasis
    Repeats letters for expressive emphasis (e.g., soooo).

  18. Lowercase
    Converts all text to lowercase.

  19. Macron / Diacritic Styling
    Adds diacritics uncommon in modern English.

  20. OCR Errors
    Introduces character confusions typical of optical character recognition.

  21. Orthographic Errors
    Applies plausible spelling mistakes.

  22. Scripted Text
    Uses decorative or stylized Unicode script characters.

  23. Similar Words
    Substitutes near-synonyms or easily confusable words.

  24. Space Removal
    Removes spaces between words.

  25. Spaced Styling
    Inserts extra spacing between characters or words.

  26. Spelled-Out Forms
    Replaces numerals or symbols with fully spelled-out equivalents.

  27. Superscript / Subscript Styling
    Uses Unicode superscripts or subscripts.

  28. Web Search Query
    Rewrites questions in keyword-heavy search-engine style.

  29. Word Reordering
    Alters word order while preserving overall meaning.


Annotations

Annotation Process

All canonical questions and perturbations were manually created and reviewed by the TokSuite team. Perturbations were designed to reflect realistic surface-form variations encountered in English text processing.

Annotators

Researchers and contributors involved in the TokSuite project with expertise in NLP and tokenizer behavior.


Personal and Sensitive Information

The dataset contains no personal, sensitive, or identifying information. All questions are general-knowledge based.


Considerations for Using the Data

Social Impact

This dataset supports research into more robust and equitable language technologies by highlighting how tokenization choices affect model behavior, even in high-resource languages like English.

Biases and Limitations

  • Focuses on Standard English
  • Evaluation-only dataset with limited size
  • Multiple-choice format
  • Does not cover domain-specific or conversational tasks

Additional Information

Dataset Curators

TokSuite Research Team (R3).

Licensing

MIT License.

Citation

If you use this dataset, please cite the TokSuite paper:

@inproceedings{toksuite2026,
  title={TokSuite: Measuring the Impact of Tokenizer Choice on Language Model Behavior},
  author={Altıntaş, Gül Sena and Ehghaghi, Malikeh and Lester, Brian and Liu, Fengyuan and Zhao, Wanru and Ciccone, Marco and Raffel, Colin},
  booktitle={Preprint},
  year={2026},
  arxiv={https://arxiv.org/abs/2512.20757},
  url={TBD}
}

Paper: TokSuite: Measuring the Impact of Tokenizer Choice on Language Model Behavior

Contributions

This dataset is part of TokSuite, which includes:

  • 14 language models with identical architectures but different tokenizers
  • Multilingual benchmark datasets (English, Turkish, Italian, Farsi, Chinese)
  • Comprehensive analysis of tokenization's impact on model behavior

Contact

For questions or issues related to this dataset, please refer to the TokSuite project or contact the authors of the paper.


Part of the TokSuite Project

Understanding Tokenization's Role in Language Model Behavior

Downloads last month
557

Collection including toksuite/tokenizer_robustness_completion_english