The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed because of a cast error
Error code: DatasetGenerationCastError
Exception: DatasetGenerationCastError
Message: An error occurred while generating the dataset
All the data files must have the same columns, but at some point there are 49 new columns ({'score_cadeval', 'score_arc-ai2', 'score_frontiermath-tier-4-2025-07-01-private', 'score_bbh', 'score_mmlu', 'score_gso-bench', 'score_arc-agi', 'score_videomme', 'release_date', 'is_open_source', 'score_vpct', 'score_hle', 'score_simplebench', 'score_triviaqa', 'model_type', 'score_frontiermath-2025-02-28-private', 'context_window', 'score_gsm8k', 'score_gpqa-diamond', 'provider_slug', 'score_anli', 'score_osworld', 'score_winogrande', 'score_balrog', 'score_cybench', 'score_piqa', 'score_swe-bench-verified-bash-only', 'input_price_per_mtok', 'score_hellaswag', 'score_geobench', 'output_price_per_mtok', 'score_arc-agi-2', 'score_lech-mazur-writing', 'score_aider-polyglot', 'score_simpleqa-verified', 'score_scienceqa', 'score_fiction-livebench', 'score_otis-mock-aime-2024-2025', 'score_weirdml', 'score_deepresearch-bench', 'avg_score', 'score_apex-agents', 'score_chess-puzzles', 'score_lambada', 'score_the-agent-company', 'score_terminal-bench', 'score_openbookqa', 'score_math-level-5', 'provider'}) and 10 missing columns ({'category', 'models_tested', 'top_score_1', 'top_model_1', 'unit', 'max_score', 'top_model_2', 'top_score_3', 'top_score_2', 'top_model_3'}).
This happened while the csv dataset builder was generating data using
hf://datasets/DropTheHQ/benchgecko-ai-models/models.csv (at revision ef51a045131e9855602e4e25afd7a94e86ec91a1), [/tmp/hf-datasets-cache/medium/datasets/66639039558395-config-parquet-and-info-DropTheHQ-benchgecko-ai-m-5349339c/hub/datasets--DropTheHQ--benchgecko-ai-models/snapshots/ef51a045131e9855602e4e25afd7a94e86ec91a1/benchmarks.csv (origin=hf://datasets/DropTheHQ/benchgecko-ai-models@ef51a045131e9855602e4e25afd7a94e86ec91a1/benchmarks.csv), /tmp/hf-datasets-cache/medium/datasets/66639039558395-config-parquet-and-info-DropTheHQ-benchgecko-ai-m-5349339c/hub/datasets--DropTheHQ--benchgecko-ai-models/snapshots/ef51a045131e9855602e4e25afd7a94e86ec91a1/models.csv (origin=hf://datasets/DropTheHQ/benchgecko-ai-models@ef51a045131e9855602e4e25afd7a94e86ec91a1/models.csv), /tmp/hf-datasets-cache/medium/datasets/66639039558395-config-parquet-and-info-DropTheHQ-benchgecko-ai-m-5349339c/hub/datasets--DropTheHQ--benchgecko-ai-models/snapshots/ef51a045131e9855602e4e25afd7a94e86ec91a1/providers.csv (origin=hf://datasets/DropTheHQ/benchgecko-ai-models@ef51a045131e9855602e4e25afd7a94e86ec91a1/providers.csv), /tmp/hf-datasets-cache/medium/datasets/66639039558395-config-parquet-and-info-DropTheHQ-benchgecko-ai-m-5349339c/hub/datasets--DropTheHQ--benchgecko-ai-models/snapshots/ef51a045131e9855602e4e25afd7a94e86ec91a1/scores.csv (origin=hf://datasets/DropTheHQ/benchgecko-ai-models@ef51a045131e9855602e4e25afd7a94e86ec91a1/scores.csv)]
Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)
Traceback: Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1890, in _prepare_split_single
writer.write_table(table)
File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 760, in write_table
pa_table = table_cast(pa_table, self._schema)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2272, in table_cast
return cast_table_to_schema(table, schema)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2218, in cast_table_to_schema
raise CastError(
datasets.table.CastError: Couldn't cast
slug: string
name: string
provider: string
provider_slug: string
avg_score: double
input_price_per_mtok: double
output_price_per_mtok: double
context_window: int64
is_open_source: bool
release_date: string
model_type: string
score_aider-polyglot: double
score_anli: double
score_apex-agents: double
score_arc-agi: double
score_arc-agi-2: double
score_arc-ai2: double
score_balrog: double
score_bbh: double
score_cadeval: double
score_chess-puzzles: double
score_cybench: double
score_deepresearch-bench: double
score_fiction-livebench: double
score_frontiermath-2025-02-28-private: double
score_frontiermath-tier-4-2025-07-01-private: double
score_geobench: double
score_gpqa-diamond: double
score_gsm8k: double
score_gso-bench: double
score_hellaswag: double
score_hle: double
score_lambada: double
score_lech-mazur-writing: double
score_math-level-5: double
score_mmlu: double
score_openbookqa: double
score_osworld: double
score_otis-mock-aime-2024-2025: double
score_piqa: double
score_scienceqa: double
score_simplebench: double
score_simpleqa-verified: double
score_swe-bench-verified-bash-only: double
score_terminal-bench: double
score_the-agent-company: double
score_triviaqa: double
score_videomme: double
score_vpct: double
score_weirdml: double
score_winogrande: double
-- schema metadata --
pandas: '{"index_columns": [{"kind": "range", "name": null, "start": 0, "' + 7048
to
{'slug': Value('string'), 'name': Value('string'), 'category': Value('string'), 'max_score': Value('int64'), 'unit': Value('string'), 'models_tested': Value('int64'), 'top_model_1': Value('string'), 'top_score_1': Value('float64'), 'top_model_2': Value('string'), 'top_score_2': Value('float64'), 'top_model_3': Value('string'), 'top_score_3': Value('float64')}
because column names don't match
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1347, in compute_config_parquet_and_info_response
parquet_operations = convert_to_parquet(builder)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 980, in convert_to_parquet
builder.download_and_prepare(
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 884, in download_and_prepare
self._download_and_prepare(
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 947, in _download_and_prepare
self._prepare_split(split_generator, **prepare_split_kwargs)
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1739, in _prepare_split
for job_id, done, content in self._prepare_split_single(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1892, in _prepare_split_single
raise DatasetGenerationCastError.from_cast_error(
datasets.exceptions.DatasetGenerationCastError: An error occurred while generating the dataset
All the data files must have the same columns, but at some point there are 49 new columns ({'score_cadeval', 'score_arc-ai2', 'score_frontiermath-tier-4-2025-07-01-private', 'score_bbh', 'score_mmlu', 'score_gso-bench', 'score_arc-agi', 'score_videomme', 'release_date', 'is_open_source', 'score_vpct', 'score_hle', 'score_simplebench', 'score_triviaqa', 'model_type', 'score_frontiermath-2025-02-28-private', 'context_window', 'score_gsm8k', 'score_gpqa-diamond', 'provider_slug', 'score_anli', 'score_osworld', 'score_winogrande', 'score_balrog', 'score_cybench', 'score_piqa', 'score_swe-bench-verified-bash-only', 'input_price_per_mtok', 'score_hellaswag', 'score_geobench', 'output_price_per_mtok', 'score_arc-agi-2', 'score_lech-mazur-writing', 'score_aider-polyglot', 'score_simpleqa-verified', 'score_scienceqa', 'score_fiction-livebench', 'score_otis-mock-aime-2024-2025', 'score_weirdml', 'score_deepresearch-bench', 'avg_score', 'score_apex-agents', 'score_chess-puzzles', 'score_lambada', 'score_the-agent-company', 'score_terminal-bench', 'score_openbookqa', 'score_math-level-5', 'provider'}) and 10 missing columns ({'category', 'models_tested', 'top_score_1', 'top_model_1', 'unit', 'max_score', 'top_model_2', 'top_score_3', 'top_score_2', 'top_model_3'}).
This happened while the csv dataset builder was generating data using
hf://datasets/DropTheHQ/benchgecko-ai-models/models.csv (at revision ef51a045131e9855602e4e25afd7a94e86ec91a1), [/tmp/hf-datasets-cache/medium/datasets/66639039558395-config-parquet-and-info-DropTheHQ-benchgecko-ai-m-5349339c/hub/datasets--DropTheHQ--benchgecko-ai-models/snapshots/ef51a045131e9855602e4e25afd7a94e86ec91a1/benchmarks.csv (origin=hf://datasets/DropTheHQ/benchgecko-ai-models@ef51a045131e9855602e4e25afd7a94e86ec91a1/benchmarks.csv), /tmp/hf-datasets-cache/medium/datasets/66639039558395-config-parquet-and-info-DropTheHQ-benchgecko-ai-m-5349339c/hub/datasets--DropTheHQ--benchgecko-ai-models/snapshots/ef51a045131e9855602e4e25afd7a94e86ec91a1/models.csv (origin=hf://datasets/DropTheHQ/benchgecko-ai-models@ef51a045131e9855602e4e25afd7a94e86ec91a1/models.csv), /tmp/hf-datasets-cache/medium/datasets/66639039558395-config-parquet-and-info-DropTheHQ-benchgecko-ai-m-5349339c/hub/datasets--DropTheHQ--benchgecko-ai-models/snapshots/ef51a045131e9855602e4e25afd7a94e86ec91a1/providers.csv (origin=hf://datasets/DropTheHQ/benchgecko-ai-models@ef51a045131e9855602e4e25afd7a94e86ec91a1/providers.csv), /tmp/hf-datasets-cache/medium/datasets/66639039558395-config-parquet-and-info-DropTheHQ-benchgecko-ai-m-5349339c/hub/datasets--DropTheHQ--benchgecko-ai-models/snapshots/ef51a045131e9855602e4e25afd7a94e86ec91a1/scores.csv (origin=hf://datasets/DropTheHQ/benchgecko-ai-models@ef51a045131e9855602e4e25afd7a94e86ec91a1/scores.csv)]
Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
slug string | name string | category string | max_score int64 | unit string | models_tested int64 | top_model_1 string | top_score_1 float64 | top_model_2 string | top_score_2 float64 | top_model_3 string | top_score_3 float64 |
|---|---|---|---|---|---|---|---|---|---|---|---|
gpqa-diamond | GPQA diamond | knowledge | 100 | % | 115 | Gemini 3.1 Pro Preview | 92.13 | GPT-5.4 | 91.07 | Gemini 3 Pro | 90.15 |
otis-mock-aime-2024-2025 | OTIS Mock AIME 2024-2025 | math | 100 | % | 105 | GPT-5.2 Chat | 96.11 | GPT-5.2 | 96.11 | Gemini 3.1 Pro Preview | 95.6 |
mmlu | MMLU | knowledge | 100 | % | 92 | GPT-4o (2024-11-20) | 84.13 | DeepSeek V3 | 82.93 | Gemini 1.5 Pro (Sept 2024) | 82.53 |
math-level-5 | MATH level 5 | math | 100 | % | 89 | GPT-5 Chat | 98.13 | GPT-5 | 98.13 | GPT-5 Mini | 97.85 |
weirdml | WeirdML | coding | 100 | % | 87 | Claude Opus 4.6 | 77.9 | GPT-5.2 Chat | 72.2 | GPT-5.2 | 72.2 |
simplebench | SimpleBench | reasoning | 100 | % | 61 | Gemini 3.1 Pro Preview | 75.52 | Gemini 3 Pro | 71.68 | GPT-5.4 Pro | 68.92 |
frontiermath-2025-02-28-private | FrontierMath-2025-02-28-Private | math | 100 | % | 60 | GPT-5.4 Pro | 50 | GPT-5.4 | 47.6 | Claude Opus 4.6 | 40.7 |
aider-polyglot | Aider polyglot | coding | 100 | % | 55 | GPT-5 Chat | 88 | GPT-5 | 88 | o3 Pro | 84.9 |
fiction-livebench | Fiction.LiveBench | knowledge | 100 | % | 53 | GPT-5 Chat | 97.2 | GPT-5 | 97.2 | o3 Pro | 97.2 |
arc-agi-2 | ARC-AGI-2 | reasoning | 100 | % | 52 | GPT-5.4 Pro | 83.33 | Gemini 3.1 Pro Preview | 77.1 | GPT-5.4 | 73.95 |
lech-mazur-writing | Lech Mazur Writing | knowledge | 100 | % | 49 | Kimi K2 0905 | 87.29 | GPT-5 Chat | 87.23 | GPT-5 | 87.23 |
arc-ai2 | ARC AI2 | knowledge | 100 | % | 48 | DeepSeek V3 | 93.73 | Llama 3.1-405B | 93.73 | Qwen2.5 72B Instruct | 92.67 |
gsm8k | GSM8K | math | 100 | % | 48 | GPT-4o-mini (2024-07-18) | 91.3 | GPT-4o-mini | 91.3 | Qwen2.5 Coder 32B Instruct | 91.1 |
winogrande | Winogrande | knowledge | 100 | % | 47 | Llama 3.1-405B | 78.4 | Claude 3 Opus | 77 | Falcon-180B | 74.2 |
frontiermath-tier-4-2025-07-01-private | FrontierMath-Tier-4-2025-07-01-Private | math | 100 | % | 39 | GPT-5.4 Pro | 37.5 | GPT-5.4 | 27.1 | Claude Opus 4.6 | 22.9 |
bbh | BBH | reasoning | 100 | % | 37 | DeepSeek V3 | 83.33 | Llama 3.1-405B | 77.2 | phi-3-medium 14B | 75.2 |
hellaswag | HellaSwag | knowledge | 100 | % | 37 | Llama 3.1-405B | 85.6 | Falcon-180B | 85.33 | DeepSeek V3 | 85.2 |
arc-agi | ARC-AGI | reasoning | 100 | % | 37 | Gemini 3.1 Pro Preview | 98 | Claude Opus 4.6 | 94 | GPT-5.2 Chat | 86.2 |
simpleqa-verified | SimpleQA Verified | knowledge | 100 | % | 36 | Gemini 3.1 Pro Preview | 77.3 | Gemini 3 Pro | 72.9 | Qwen3 Max | 67.47 |
piqa | PIQA | knowledge | 100 | % | 36 | GPT-4o-mini (2024-07-18) | 77.4 | GPT-4o-mini | 77.4 | Gemini 1.5 Flash (Sep 2024) | 75 |
swe-bench-verified-bash-only | SWE-Bench Verified (Bash Only) | coding | 100 | % | 32 | Claude Opus 4.5 | 74.4 | Gemini 3 Pro | 74.2 | GPT-5.2 Chat | 71.8 |
triviaqa | TriviaQA | knowledge | 100 | % | 31 | Llama 2-70B | 87.6 | Claude 2 | 87.5 | LLaMA-65B | 86 |
chess-puzzles | Chess Puzzles | knowledge | 100 | % | 29 | Gemini 3.1 Pro Preview | 55 | GPT-5.2 Chat | 49 | GPT-5.2 | 49 |
geobench | GeoBench | knowledge | 100 | % | 29 | Gemini 3 Flash Preview | 88 | Gemini 3 Pro | 84 | GPT-5 Chat | 81 |
terminal-bench | Terminal Bench | coding | 100 | % | 27 | Gemini 3.1 Pro Preview | 78.4 | Claude Opus 4.6 | 69.9 | GPT-5.2 Chat | 64.9 |
hle | HLE | knowledge | 100 | % | 27 | Gemini 3 Pro | 34.37 | Claude Opus 4.6 | 31.13 | GPT-5 Pro | 28.19 |
openbookqa | OpenBookQA | knowledge | 100 | % | 27 | phi-3-mini 3.8B | 84 | phi-3-small 7.4B | 84 | phi-3-medium 14B | 83.2 |
vpct | VPCT | knowledge | 100 | % | 26 | Gemini 3 Pro | 86.5 | GPT-5.2 Chat | 76 | GPT-5.2 | 76 |
gso-bench | GSO-Bench | coding | 100 | % | 23 | Claude Opus 4.6 | 33.33 | GPT-5.2 Chat | 27.4 | GPT-5.2 | 27.4 |
apex-agents | APEX-Agents | agentic | 100 | % | 21 | GPT-5.4 | 35.9 | GPT-5.2 Chat | 34.3 | GPT-5.2 | 34.3 |
balrog | Balrog | knowledge | 100 | % | 20 | Gemini 3 Flash Preview | 48.1 | Grok 4 | 43.6 | DeepSeek-R1 | 34.9 |
cybench | Cybench | coding | 100 | % | 17 | Claude Sonnet 4.5 | 55 | Claude Opus 4.1 | 38 | Claude Opus 4 | 38 |
lambada | LAMBADA | knowledge | 100 | % | 16 | Falcon-180B | 79.8 | Llama 2-70B | 78.9 | LLaMA-65B | 77.7 |
cadeval | CadEval | coding | 100 | % | 15 | o3 | 74 | o4 Mini | 62 | o1 | 56 |
deepresearch-bench | DeepResearch Bench | knowledge | 100 | % | 12 | Claude Sonnet 4.5 | 52.6 | GPT-5 Chat | 51 | GPT-5 | 51 |
videomme | VideoMME | multimodal | 100 | % | 11 | Gemini 1.5 Pro (Feb 2024) | 66.67 | Qwen2.5 72B Instruct | 64.67 | GPT-4o (2024-11-20) | 62.53 |
the-agent-company | The Agent Company | agentic | 100 | % | 10 | DeepSeek V3.2 Exp | 42.9 | Claude Sonnet 4 | 33.1 | Claude 3.7 Sonnet | 30.9 |
osworld | OSWorld | agentic | 100 | % | 8 | Claude Opus 4.5 | 66.3 | Kimi K2.5 | 63.3 | Claude Sonnet 4.5 | 62.9 |
anli | ANLI | knowledge | 100 | % | 8 | phi-3-small 7.4B | 37.15 | Llama 3 8B Instruct | 35.95 | phi-3-medium 14B | 33.7 |
scienceqa | ScienceQA | knowledge | 100 | % | 5 | Claude 3 Haiku | 62.67 | Llama 2-13B | 41.04 | LLaMA-13B | 24.44 |
claude-instant | Claude Instant | null | null | null | null | null | null | null | null | null | null |
deepseek-v2-moe-236b-may-2024 | DeepSeek-V2 (MoE-236B, May 2024) | null | null | null | null | null | null | null | null | null | null |
qwen-2-5-coder-32b-instruct | Qwen2.5 Coder 32B Instruct | null | null | null | null | null | null | null | null | null | null |
stable-beluga-2 | Stable Beluga 2 | null | null | null | null | null | null | null | null | null | null |
gemini-3-1-pro-preview | Gemini 3.1 Pro Preview | null | null | null | null | null | null | null | null | null | null |
phi-3-small-7-4b | phi-3-small 7.4B | null | null | null | null | null | null | null | null | null | null |
falcon-180b | Falcon-180B | null | null | null | null | null | null | null | null | null | null |
qwen3-max | Qwen3 Max | null | null | null | null | null | null | null | null | null | null |
o3-pro | o3 Pro | null | null | null | null | null | null | null | null | null | null |
kimi-k2-0905 | Kimi K2 0905 | null | null | null | null | null | null | null | null | null | null |
llama-65b | LLaMA-65B | null | null | null | null | null | null | null | null | null | null |
qwen2-5-coder-7b-instruct | Qwen2.5 Coder 7B Instruct | null | null | null | null | null | null | null | null | null | null |
qwen-14b | Qwen-14B | null | null | null | null | null | null | null | null | null | null |
phi-3-mini-3-8b | phi-3-mini 3.8B | null | null | null | null | null | null | null | null | null | null |
deepseek-v3-2-exp | DeepSeek V3.2 Exp | null | null | null | null | null | null | null | null | null | null |
gpt-5-4-pro | GPT-5.4 Pro | null | null | null | null | null | null | null | null | null | null |
gpt-5-4 | GPT-5.4 | null | null | null | null | null | null | null | null | null | null |
phi-3-medium-14b | phi-3-medium 14B | null | null | null | null | null | null | null | null | null | null |
falcon-2-11b | Falcon 2 11B | null | null | null | null | null | null | null | null | null | null |
llama-33b | LLaMA-33B | null | null | null | null | null | null | null | null | null | null |
mixtral-8x7b-instruct | Mixtral 8x7B Instruct | null | null | null | null | null | null | null | null | null | null |
mistral-7b-v0-1 | Mistral 7B v0.1 | null | null | null | null | null | null | null | null | null | null |
gemma-7b | Gemma 7B | null | null | null | null | null | null | null | null | null | null |
deepseek-chat | DeepSeek V3 | null | null | null | null | null | null | null | null | null | null |
gemini-3-pro | Gemini 3 Pro | null | null | null | null | null | null | null | null | null | null |
llama-2-70b | Llama 2-70B | null | null | null | null | null | null | null | null | null | null |
qwen3-235b-a22b | Qwen3 235B A22B | null | null | null | null | null | null | null | null | null | null |
grok-4-fast | Grok 4 Fast | null | null | null | null | null | null | null | null | null | null |
claude-opus-4-6 | Claude Opus 4.6 | null | null | null | null | null | null | null | null | null | null |
gpt-5-2-chat | GPT-5.2 Chat | null | null | null | null | null | null | null | null | null | null |
gpt-5-2 | GPT-5.2 | null | null | null | null | null | null | null | null | null | null |
o1 | o1 | null | null | null | null | null | null | null | null | null | null |
falcon-40b | Falcon-40B | null | null | null | null | null | null | null | null | null | null |
gpt-5-chat | GPT-5 Chat | null | null | null | null | null | null | null | null | null | null |
gpt-5 | GPT-5 | null | null | null | null | null | null | null | null | null | null |
llama-2-13b | Llama 2-13B | null | null | null | null | null | null | null | null | null | null |
gemini-2-0-pro | Gemini 2.0 Pro | null | null | null | null | null | null | null | null | null | null |
nemotron-4-15b | Nemotron-4 15B | null | null | null | null | null | null | null | null | null | null |
yi-34b | Yi-34B | null | null | null | null | null | null | null | null | null | null |
qwen-2-5-72b-instruct | Qwen2.5 72B Instruct | null | null | null | null | null | null | null | null | null | null |
o3 | o3 | null | null | null | null | null | null | null | null | null | null |
deepseek-chat-v3-1 | DeepSeek V3.1 | null | null | null | null | null | null | null | null | null | null |
gemini-3-flash-preview | Gemini 3 Flash Preview | null | null | null | null | null | null | null | null | null | null |
llama-2-34b | Llama 2-34B | null | null | null | null | null | null | null | null | null | null |
gemini-2-0-flash-001 | Gemini 2.0 Flash | null | null | null | null | null | null | null | null | null | null |
llama-3-1-405b | Llama 3.1-405B | null | null | null | null | null | null | null | null | null | null |
qwen-7b | Qwen-7B | null | null | null | null | null | null | null | null | null | null |
deepseek-r1-may-2025 | DeepSeek-R1 (May 2025) | null | null | null | null | null | null | null | null | null | null |
baichuan2-13b | Baichuan2-13B | null | null | null | null | null | null | null | null | null | null |
qwen3-235b-a22b-thinking-2507 | Qwen3 235B A22B Thinking 2507 | null | null | null | null | null | null | null | null | null | null |
grok-4 | Grok 4 | null | null | null | null | null | null | null | null | null | null |
deepseek-r1 | DeepSeek-R1 | null | null | null | null | null | null | null | null | null | null |
claude-sonnet-4-6 | Claude Sonnet 4.6 | null | null | null | null | null | null | null | null | null | null |
grok-3-mini | Grok 3 Mini | null | null | null | null | null | null | null | null | null | null |
grok-3-mini-beta | Grok 3 Mini Beta | null | null | null | null | null | null | null | null | null | null |
claude-opus-4-5 | Claude Opus 4.5 | null | null | null | null | null | null | null | null | null | null |
o4-mini | o4 Mini | null | null | null | null | null | null | null | null | null | null |
phi-4 | Phi 4 | null | null | null | null | null | null | null | null | null | null |
llama-13b | LLaMA-13B | null | null | null | null | null | null | null | null | null | null |
gpt-4-turbo | GPT-4 Turbo | null | null | null | null | null | null | null | null | null | null |
End of preview.