_leaderboard stringclasses 1
value | _developer stringclasses 559
values | _model stringlengths 9 102 | _uuid stringlengths 36 36 | schema_version stringclasses 1
value | evaluation_id stringlengths 35 133 | retrieved_timestamp stringlengths 13 18 | source_data stringclasses 1
value | evaluation_source_name stringclasses 1
value | evaluation_source_type stringclasses 1
value | source_organization_name stringclasses 1
value | source_organization_url null | source_organization_logo_url null | evaluator_relationship stringclasses 1
value | model_name stringlengths 4 102 | model_id stringlengths 9 102 | model_developer stringclasses 559
values | model_inference_platform stringclasses 1
value | evaluation_results stringlengths 1.35k 1.41k | additional_details stringclasses 660
values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-8B | d25510e4-6549-4f64-8ec4-37ac8671050c | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-8B/1762652580.237391 | 1762652580.237392 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-8B | jaspionjader/Kosmos-EVAA-gamma-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45722460848326885}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-6 | 0e1cd676-f95b-4562-8c5d-e932f148dc23 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-6/1762652580.276035 | 1762652580.276036 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-6 | jaspionjader/slu-6 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.41166216749336204}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-54 | 9a2d7235-84cf-43f6-8855-68d0bf85e6e3 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-54/1762652580.258788 | 1762652580.258792 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-54 | jaspionjader/bh-54 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48405231452545916}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/kstc-1-8b | cd7e14cb-b1f1-47d8-81a9-960da8ac4e05 | 0.0.1 | hfopenllm_v2/jaspionjader_kstc-1-8b/1762652580.2702851 | 1762652580.270286 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/kstc-1-8b | jaspionjader/kstc-1-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4642936297911763}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/kstc-11-8b | 41b46842-dffa-4791-8225-99d676f563c9 | 0.0.1 | hfopenllm_v2/jaspionjader_kstc-11-8b/1762652580.270522 | 1762652580.270522 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/kstc-11-8b | jaspionjader/kstc-11-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4757343847657549}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-7 | b63d1462-f84b-4d20-86d6-1a54cf4eb81f | 0.0.1 | hfopenllm_v2/jaspionjader_bh-7/1762652580.261788 | 1762652580.261791 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-7 | jaspionjader/bh-7 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4623953332712758}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-TSN-light-8B | 9f0aa20f-8687-4c21-b222-39a322f90842 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-TSN-light-8B/1762652580.236298 | 1762652580.236299 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-TSN-light-8B | jaspionjader/Kosmos-EVAA-TSN-light-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46849027247702757}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v2-8B | 8f16aed2-8b31-48cc-b874-8d437f26f3db | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v2-8B/1762652580.241379 | 1762652580.2413802 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v2-8B | jaspionjader/Kosmos-EVAA-v2-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4395891789341171}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-v17-8B | bd4cc259-d535-437a-afc5-d74a60154b07 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-v17-8B/1762652580.239734 | 1762652580.239739 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-v17-8B | jaspionjader/Kosmos-EVAA-gamma-v17-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4462337708391512}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v10-8B | c57d95da-1b6f-4ce7-8c42-f1129fc1e55e | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v10-8B/1762652580.2406652 | 1762652580.2406662 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v10-8B | jaspionjader/Kosmos-EVAA-v10-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4261503920708165}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-32 | 1095577f-7b50-4854-9c7c-5beb59206e60 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-32/1762652580.274382 | 1762652580.274383 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-32 | jaspionjader/slu-32 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45155409868039026}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-32 | a56c62cc-c318-4de4-b6c7-0fa10229a127 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-32/1762652580.253373 | 1762652580.2533739 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-32 | jaspionjader/bh-32 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4635943740386619}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-29 | 012eeeed-c556-460d-82f6-34bdc31da5cf | 0.0.1 | hfopenllm_v2/jaspionjader_bh-29/1762652580.252519 | 1762652580.2525198 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-29 | jaspionjader/bh-29 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46881496651107946}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bbb-3 | 94668ddb-d2fb-44e2-8ed7-10179d145366 | 0.0.1 | hfopenllm_v2/jaspionjader_bbb-3/1762652580.24635 | 1762652580.246351 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bbb-3 | jaspionjader/bbb-3 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.416832892281369}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-64 | 90830134-43d5-4d0c-9a93-4be2c1c7dba8 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-64/1762652580.261374 | 1762652580.261375 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-64 | jaspionjader/bh-64 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.41401038134372353}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-13 | 1a1eaa84-9926-4c4b-b254-96cd667c25ac | 0.0.1 | hfopenllm_v2/jaspionjader_slu-13/1762652580.272234 | 1762652580.272237 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-13 | jaspionjader/slu-13 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4378404854674486}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bbb-6 | ef8025de-fe9f-4a79-97f6-c26c18ab049a | 0.0.1 | hfopenllm_v2/jaspionjader_bbb-6/1762652580.247051 | 1762652580.247051 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bbb-6 | jaspionjader/bbb-6 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48797413086166924}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-47 | 9f30c4d4-4a3c-459e-8444-e143ef75f84e | 0.0.1 | hfopenllm_v2/jaspionjader_bh-47/1762652580.256935 | 1762652580.2569358 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-47 | jaspionjader/bh-47 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46516797652451053}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-31 | a637936e-646b-4c21-964a-61e253fd3705 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-31/1762652580.253162 | 1762652580.253163 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-31 | jaspionjader/bh-31 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4727367828472896}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-11 | 0091eabc-3888-4e1a-a29d-8c4e98b599f2 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-11/1762652580.272018 | 1762652580.272018 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-11 | jaspionjader/slu-11 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.372519359743259}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-41 | db0c4182-7391-40e7-ad6e-5374c8eb28e1 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-41/1762652580.2558541 | 1762652580.2558541 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-41 | jaspionjader/bh-41 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4739856912990864}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-62 | 85bd08bf-bdc3-42fb-b8f9-3d83e32921bc | 0.0.1 | hfopenllm_v2/jaspionjader_bh-62/1762652580.260948 | 1762652580.260949 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-62 | jaspionjader/bh-62 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.41498446344587914}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-Aurora_faustus-8B | 0563ee22-d981-45cb-83f8-7dbdb2734d10 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-Aurora_faustus-8B/1762652580.231864 | 1762652580.2318652 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-Aurora_faustus-8B | jaspionjader/Kosmos-Aurora_faustus-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.443236168920686}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/gamma-Kosmos-EVAA-8B | 6e5584a8-5b8e-48ce-8b80-2d39a74a9b0d | 0.0.1 | hfopenllm_v2/jaspionjader_gamma-Kosmos-EVAA-8B/1762652580.268576 | 1762652580.268577 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/gamma-Kosmos-EVAA-8B | jaspionjader/gamma-Kosmos-EVAA-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42500121898784116}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v34-8B | 11486e0e-a9e3-43b0-b26e-299a86555d16 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v34-8B/1762652580.235871 | 1762652580.235871 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v34-8B | jaspionjader/Kosmos-EVAA-PRP-v34-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45625052638111324}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/f-7-8b | e8c5d934-c9b6-460c-bd45-c4a3e2d26bed | 0.0.1 | hfopenllm_v2/jaspionjader_f-7-8b/1762652580.2664478 | 1762652580.266449 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/f-7-8b | jaspionjader/f-7-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4462337708391512}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-10 | 2bcc7f9a-9c36-487e-8522-bfbe1910b857 | 0.0.1 | hfopenllm_v2/jaspionjader_test-10/1762652580.2777631 | 1762652580.277764 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-10 | jaspionjader/test-10 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4578241288669619}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-v15-8B | 24e11e0c-fb61-46c1-a05e-c533eb392195 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-v15-8B/1762652580.239064 | 1762652580.2390652 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-v15-8B | jaspionjader/Kosmos-EVAA-gamma-v15-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4654428028741517}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-37 | e64e5fe0-c726-4b9d-9d7b-952e7c7508ab | 0.0.1 | hfopenllm_v2/jaspionjader_slu-37/1762652580.2757561 | 1762652580.275757 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-37 | jaspionjader/slu-37 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4533526598314694}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/fct-9-8b | 4d1ddf64-4626-4877-a0fa-84e06f6cf977 | 0.0.1 | hfopenllm_v2/jaspionjader_fct-9-8b/1762652580.267691 | 1762652580.267692 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/fct-9-8b | jaspionjader/fct-9-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4353925362482657}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v23-8B | 5d5ae047-72d1-4083-8e28-dcce7337ed25 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v23-8B/1762652580.233495 | 1762652580.233495 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v23-8B | jaspionjader/Kosmos-EVAA-PRP-v23-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4040933611705829}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/f-8-8b | dad898e1-ee18-4864-b432-462d17ac8006 | 0.0.1 | hfopenllm_v2/jaspionjader_f-8-8b/1762652580.266931 | 1762652580.266932 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/f-8-8b | jaspionjader/f-8-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4739358236146758}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Auro-Kosmos-EVAA-v2.3-8B | 9c7ee100-754e-4665-8527-452021a2243b | 0.0.1 | hfopenllm_v2/jaspionjader_Auro-Kosmos-EVAA-v2.3-8B/1762652580.231667 | 1762652580.231667 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Auro-Kosmos-EVAA-v2.3-8B | jaspionjader/Auro-Kosmos-EVAA-v2.3-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42712447417297217}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-61 | 00b1b367-c4eb-4048-b80d-a8253e7c2048 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-61/1762652580.260743 | 1762652580.260743 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-61 | jaspionjader/bh-61 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42467652495378927}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-18 | 6aad7ade-7bd0-4515-b4ac-2299c58da098 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-18/1762652580.249514 | 1762652580.249515 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-18 | jaspionjader/bh-18 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47246195649764844}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-59 | 974b1542-8716-4ea3-b097-f9893c9c9656 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-59/1762652580.260088 | 1762652580.2600892 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-59 | jaspionjader/bh-59 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43414362779646887}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v25-8B | 81c8704c-7124-42d1-b320-77e31e35898b | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v25-8B/1762652580.23391 | 1762652580.23391 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v25-8B | jaspionjader/Kosmos-EVAA-PRP-v25-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4420869958377106}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/knf-2-8b | 267e641c-7fbd-40d3-a9b7-eb3621240b2a | 0.0.1 | hfopenllm_v2/jaspionjader_knf-2-8b/1762652580.269415 | 1762652580.2694159 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/knf-2-8b | jaspionjader/knf-2-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42500121898784116}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/gamma-Kosmos-EVAA-v3-8B | d461545f-ebcb-49e2-94ce-a6591e31a94a | 0.0.1 | hfopenllm_v2/jaspionjader_gamma-Kosmos-EVAA-v3-8B/1762652580.269119 | 1762652580.26912 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/gamma-Kosmos-EVAA-v3-8B | jaspionjader/gamma-Kosmos-EVAA-v3-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43326928106313467}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bbb-1 | b6ca35e1-8680-49e8-a6dd-963214be7411 | 0.0.1 | hfopenllm_v2/jaspionjader_bbb-1/1762652580.2456498 | 1762652580.245653 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bbb-1 | jaspionjader/bbb-1 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4864005283758206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/knfp-2-8b | 0bd6a333-afc0-43a4-9d14-fa44c2364184 | 0.0.1 | hfopenllm_v2/jaspionjader_knfp-2-8b/1762652580.2696629 | 1762652580.269664 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/knfp-2-8b | jaspionjader/knfp-2-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5327120928026525}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/TSN-Kosmos-EVAA-v2-8B | 2ce2b8e4-0cd4-4001-8790-ad5e26e3e45c | 0.0.1 | hfopenllm_v2/jaspionjader_TSN-Kosmos-EVAA-v2-8B/1762652580.2454138 | 1762652580.245415 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/TSN-Kosmos-EVAA-v2-8B | jaspionjader/TSN-Kosmos-EVAA-v2-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46669171132594844}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-24 | 0b27b829-6588-4f7b-80fe-6e6767287a38 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-24/1762652580.251392 | 1762652580.251392 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-24 | jaspionjader/bh-24 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4715377420799035}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-26 | 0218b7de-bbd7-4196-8fec-3f6fb790a3c1 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-26/1762652580.251851 | 1762652580.251852 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-26 | jaspionjader/bh-26 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4690897928607206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/knfp-3-8b | 38a5c599-a098-42f4-a7cb-acee487e382a | 0.0.1 | hfopenllm_v2/jaspionjader_knfp-3-8b/1762652580.2700531 | 1762652580.2700539 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/knfp-3-8b | jaspionjader/knfp-3-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49456885508229276}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v29-8B | 86e94a19-e497-4539-802b-597ce0e0ced0 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v29-8B/1762652580.234771 | 1762652580.234771 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v29-8B | jaspionjader/Kosmos-EVAA-PRP-v29-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4487315877427448}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-TSN-v21-8B | d9c819c2-a3f6-481e-bd71-47912aef9847 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-TSN-v21-8B/1762652580.2369542 | 1762652580.236955 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-TSN-v21-8B | jaspionjader/Kosmos-EVAA-TSN-v21-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46701640536000033}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-12 | 736ee66e-bd19-4275-afaf-73c2112c2fbd | 0.0.1 | hfopenllm_v2/jaspionjader_bh-12/1762652580.248367 | 1762652580.248368 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-12 | jaspionjader/bh-12 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47338617091539337}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-34 | 050afa51-be7c-4cad-ae8b-bd63384df297 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-34/1762652580.2749598 | 1762652580.274961 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-34 | jaspionjader/slu-34 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4350678422142138}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v31-8B | 0757cecd-bc5f-4095-90ee-25920ae6670c | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v31-8B/1762652580.235214 | 1762652580.235214 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v31-8B | jaspionjader/Kosmos-EVAA-PRP-v31-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43986400528375824}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-33 | bcab8546-ea69-4207-b69b-ab982b603e55 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-33/1762652580.25359 | 1762652580.253591 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-33 | jaspionjader/bh-33 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4685401401614383}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-15 | af3bd92d-45f5-4a48-89aa-b8c956209d5a | 0.0.1 | hfopenllm_v2/jaspionjader_bh-15/1762652580.248791 | 1762652580.2487922 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-15 | jaspionjader/bh-15 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47453534399836883}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-TSN-v19-8B | 91c2897a-3ae3-402b-aadf-26d0b8d746c5 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-TSN-v19-8B/1762652580.236516 | 1762652580.2365172 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-TSN-v19-8B | jaspionjader/Kosmos-EVAA-TSN-v19-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4563502617499346}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-16 | c98928d3-0d7f-429c-927c-bf8fa432101a | 0.0.1 | hfopenllm_v2/jaspionjader_bh-16/1762652580.2489972 | 1762652580.248998 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-16 | jaspionjader/bh-16 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4730614768813415}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-63 | c9df2e30-5e2d-42cc-8597-dc354602350a | 0.0.1 | hfopenllm_v2/jaspionjader_bh-63/1762652580.261157 | 1762652580.261157 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-63 | jaspionjader/bh-63 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43077146415954115}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-49 | e574e35a-56cb-471d-b4f1-df0858f5ce66 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-49/1762652580.257362 | 1762652580.257366 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-49 | jaspionjader/bh-49 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47246195649764844}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-Fusion-8B | ac41e588-0664-44f5-9fa9-eafd6508078b | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-Fusion-8B/1762652580.23267 | 1762652580.232671 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-Fusion-8B | jaspionjader/Kosmos-EVAA-Fusion-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4417623018036587}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-Fusion-8B | 3030519e-f137-4091-9394-26a0779f0ad9 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-Fusion-8B/1762652580.2328691 | 1762652580.2328691 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-Fusion-8B | jaspionjader/Kosmos-EVAA-Fusion-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43446832183052075}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-Elusive-8b | 60d775f1-47a9-45ae-9b2f-75b95c9d96cd | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-Elusive-8b/1762652580.243371 | 1762652580.243371 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-Elusive-8b | jaspionjader/Kosmos-Elusive-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.41688275996577967}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-48 | 80bbd567-b13e-4ed4-ba85-9098639a3642 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-48/1762652580.257132 | 1762652580.257133 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-48 | jaspionjader/bh-48 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46881496651107946}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-TSN-8B | 75037d12-da94-4c55-8de5-a7cef098d4b0 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-TSN-8B/1762652580.236081 | 1762652580.2360818 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-TSN-8B | jaspionjader/Kosmos-EVAA-TSN-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47213726246359655}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-25 | 03c03447-1bf3-4721-8f9e-5ef041ab5d7d | 0.0.1 | hfopenllm_v2/jaspionjader_slu-25/1762652580.27394 | 1762652580.273941 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-25 | jaspionjader/slu-25 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4500303638789523}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-53 | 978d4a27-17c7-4f87-b3e5-27b00ffa4d80 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-53/1762652580.25855 | 1762652580.2585511 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-53 | jaspionjader/bh-53 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4779573753197073}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-22 | c0898ca4-21a7-4d83-ad2e-1aa61bd370fa | 0.0.1 | hfopenllm_v2/jaspionjader_slu-22/1762652580.2733881 | 1762652580.273391 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-22 | jaspionjader/slu-22 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4321201079801593}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v27-8B | d3dcd3f0-2f43-4b82-ba29-77a69a9b3e8f | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v27-8B/1762652580.2343428 | 1762652580.234344 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v27-8B | jaspionjader/Kosmos-EVAA-PRP-v27-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4378404854674486}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/f-6-8b | 2a71c7d7-8ae6-45e7-ab7f-54f7d31dd131 | 0.0.1 | hfopenllm_v2/jaspionjader_f-6-8b/1762652580.2661529 | 1762652580.266155 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/f-6-8b | jaspionjader/f-6-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48460196722474147}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-19 | ab7e0f6c-bca9-4f83-a4a0-5014c46e0512 | 0.0.1 | hfopenllm_v2/jaspionjader_test-19/1762652580.279826 | 1762652580.2798269 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-19 | jaspionjader/test-19 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44008896394898867}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/TSN-Kosmos-EVAA-8B | 9819f2bd-8108-4fc5-9208-ce295d860435 | 0.0.1 | hfopenllm_v2/jaspionjader_TSN-Kosmos-EVAA-8B/1762652580.2451851 | 1762652580.245186 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/TSN-Kosmos-EVAA-8B | jaspionjader/TSN-Kosmos-EVAA-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49032234471203073}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-20 | 6391f921-4de7-4e83-8bb2-8d0ef0b58d8f | 0.0.1 | hfopenllm_v2/jaspionjader_test-20/1762652580.2800388 | 1762652580.28004 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-20 | jaspionjader/test-20 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45292823042859615}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bbb-4 | 828a6bd0-a205-4327-bc77-2e8a84c0b69e | 0.0.1 | hfopenllm_v2/jaspionjader_bbb-4/1762652580.2465842 | 1762652580.2465851 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bbb-4 | jaspionjader/bbb-4 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47675833455232114}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v9-8B | cd0c4096-93ee-4a04-83b0-44063770e81b | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v9-8B/1762652580.242934 | 1762652580.242935 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v9-8B | jaspionjader/Kosmos-EVAA-v9-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43686640336529303}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/kstc-6-8b | f7d63a4b-070d-4581-acce-cd356a3dea47 | 0.0.1 | hfopenllm_v2/jaspionjader_kstc-6-8b/1762652580.2711701 | 1762652580.2711701 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/kstc-6-8b | jaspionjader/kstc-6-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49439376410147295}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-25 | a0c16d3d-e3f2-4c50-975a-70b69824b3d5 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-25/1762652580.251633 | 1762652580.251633 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-25 | jaspionjader/bh-25 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47518473206647255}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-alt-8B | 58e279d4-da0f-4e2c-a74d-c51caeaad884 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-alt-8B/1762652580.23761 | 1762652580.23761 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-alt-8B | jaspionjader/Kosmos-EVAA-gamma-alt-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4542270065648036}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-21 | a369ff4f-7fe9-4764-be74-83563dbaf635 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-21/1762652580.25052 | 1762652580.2505212 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-21 | jaspionjader/bh-21 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47001400727846554}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-light-alt-8B | abebffbf-48b5-4452-8c7a-bb1175a7e979 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-light-alt-8B/1762652580.238084 | 1762652580.238085 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-light-alt-8B | jaspionjader/Kosmos-EVAA-gamma-light-alt-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44535942410581697}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | Qwen | Qwen/Qwen1.5-0.5B-Chat | 96baee1a-7ea7-454f-ac8b-fe1bead3cd93 | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-0.5B-Chat/1762652579.835679 | 1762652579.83568 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-0.5B-Chat | Qwen/Qwen1.5-0.5B-Chat | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.18072713732895385}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.62} |
HF Open LLM v2 | Qwen | Qwen/Qwen2.5-0.5B-Instruct | 883755e2-69eb-459b-ae7f-5548914aa65e | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-0.5B-Instruct/1762652579.842189 | 1762652579.84219 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-0.5B-Instruct | Qwen/Qwen2.5-0.5B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.307122878407071}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.5} |
HF Open LLM v2 | Qwen | Qwen/Qwen2.5-0.5B-Instruct | 14d1ea99-ae05-42cd-9f2f-de1a98d9846d | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-0.5B-Instruct/1762652579.842413 | 1762652579.8424141 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-0.5B-Instruct | Qwen/Qwen2.5-0.5B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31529120511354314}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.494} |
HF Open LLM v2 | Qwen | Qwen/Qwen2.5-72B-Instruct | 9ed2a831-aa5a-4e81-b8b5-397bc8b55835 | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-72B-Instruct/1762652579.844789 | 1762652579.844789 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-72B-Instruct | Qwen/Qwen2.5-72B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.863837949972739}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706} |
HF Open LLM v2 | Qwen | Qwen/Qwen1.5-14B-Chat | e2cdcc99-a1b6-43ee-9cda-2e7ccbd0ad8d | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-14B-Chat/1762652579.837058 | 1762652579.837059 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-14B-Chat | Qwen/Qwen1.5-14B-Chat | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47680820223673187}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.167} |
HF Open LLM v2 | Qwen | Qwen/Qwen1.5-1.8B-Chat | d6107bde-875e-40f6-8471-3a3507758910 | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-1.8B-Chat/1762652579.836214 | 1762652579.836215 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-1.8B-Chat | Qwen/Qwen1.5-1.8B-Chat | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20190982149585324}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.837} |
HF Open LLM v2 | Qwen | Qwen/Qwen2.5-32B-Instruct | c921186d-6e97-46d6-b968-894159271620 | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-32B-Instruct/1762652579.843922 | 1762652579.843922 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-32B-Instruct | Qwen/Qwen2.5-32B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8346121623957765}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764} |
HF Open LLM v2 | Qwen | Qwen/Qwen2-1.5B-Instruct | 984029c7-f957-4555-8460-dfecd99f44a1 | 0.0.1 | hfopenllm_v2/Qwen_Qwen2-1.5B-Instruct/1762652579.839607 | 1762652579.839607 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2-1.5B-Instruct | Qwen/Qwen2-1.5B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3371232773485463}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544} |
HF Open LLM v2 | Qwen | Qwen/Qwen2.5-Coder-7B-Instruct | 81749833-4f2a-4883-a789-c465c11b33b6 | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-Coder-7B-Instruct/1762652579.8473449 | 1762652579.8473458 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-Coder-7B-Instruct | Qwen/Qwen2.5-Coder-7B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6147189457306613}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | Qwen | Qwen/Qwen2.5-Coder-7B-Instruct | 7629f304-5235-485b-a7f6-f5a7f91fd35c | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-Coder-7B-Instruct/1762652579.847122 | 1762652579.847123 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-Coder-7B-Instruct | Qwen/Qwen2.5-Coder-7B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6101477413263474}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | Qwen | Qwen/Qwen2.5-7B-Instruct | 7a336f2b-3b33-4fde-bce6-2d1e884a1b26 | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-7B-Instruct/1762652579.845207 | 1762652579.8452082 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-7B-Instruct | Qwen/Qwen2.5-7B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7585251576926999}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | Qwen | Qwen/Qwen2.5-Math-72B-Instruct | 64574dc3-4982-49c3-8526-09ebd5781175 | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-Math-72B-Instruct/1762652579.847774 | 1762652579.847775 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-Math-72B-Instruct | Qwen/Qwen2.5-Math-72B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4003466358151926}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706} |
HF Open LLM v2 | Qwen | Qwen/Qwen2-57B-A14B-Instruct | 50496313-dc6c-4456-8a8c-15cd8ddbb480 | 0.0.1 | hfopenllm_v2/Qwen_Qwen2-57B-A14B-Instruct/1762652579.84003 | 1762652579.840031 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2-57B-A14B-Instruct | Qwen/Qwen2-57B-A14B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6337783747124297}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2MoeForCausalLM", "params_billions": 57.409} |
HF Open LLM v2 | Qwen | Qwen/Qwen2.5-14B-Instruct | 1f3e04ab-9f97-4eda-9d40-669eda073ac3 | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-14B-Instruct/1762652579.843263 | 1762652579.843264 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-14B-Instruct | Qwen/Qwen2.5-14B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8157776920792386}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | Qwen | Qwen/Qwen1.5-7B-Chat | 42e3c9e4-bf1a-43ae-87e7-056f735abe03 | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-7B-Chat/1762652579.838321 | 1762652579.838322 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-7B-Chat | Qwen/Qwen1.5-7B-Chat | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43711574178734647}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.721} |
HF Open LLM v2 | Qwen | Qwen/Qwen2-VL-72B-Instruct | 2f749e28-b845-45ab-a628-8f9b6a9029d9 | 0.0.1 | hfopenllm_v2/Qwen_Qwen2-VL-72B-Instruct/1762652579.841569 | 1762652579.8415701 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2-VL-72B-Instruct | Qwen/Qwen2-VL-72B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5982326892644849}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2VLForConditionalGeneration", "params_billions": 73.406} |
HF Open LLM v2 | Qwen | Qwen/Qwen1.5-32B-Chat | c14a0d32-1d27-4596-90d4-10a793aef9a2 | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-32B-Chat/1762652579.8374798 | 1762652579.8374798 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-32B-Chat | Qwen/Qwen1.5-32B-Chat | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5532199009738605}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.512} |
HF Open LLM v2 | Qwen | Qwen/Qwen2.5-1.5B-Instruct | 9744dd76-a8cd-4400-92a7-f10b375710ae | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-1.5B-Instruct/1762652579.842835 | 1762652579.842836 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-1.5B-Instruct | Qwen/Qwen2.5-1.5B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4475569267321817}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.5} |
HF Open LLM v2 | Qwen | Qwen/Qwen2-72B-Instruct | d9ae7c35-ac71-4703-9cfe-bf5fb5aa688e | 0.0.1 | hfopenllm_v2/Qwen_Qwen2-72B-Instruct/1762652579.840446 | 1762652579.840447 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2-72B-Instruct | Qwen/Qwen2-72B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7989168738945996}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706} |
HF Open LLM v2 | Qwen | Qwen/Qwen1.5-110B-Chat | 7cfcae3d-b623-4cf0-9ac8-529db46d05e6 | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-110B-Chat/1762652579.836649 | 1762652579.836649 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-110B-Chat | Qwen/Qwen1.5-110B-Chat | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5938864435254014}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 111.21} |
HF Open LLM v2 | Qwen | Qwen/Qwen2-VL-7B-Instruct | 6dd0eebe-ef61-431d-bf7c-c170475bed5f | 0.0.1 | hfopenllm_v2/Qwen_Qwen2-VL-7B-Instruct/1762652579.841773 | 1762652579.841774 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2-VL-7B-Instruct | Qwen/Qwen2-VL-7B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4599218961245052}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2VLForConditionalGeneration", "params_billions": 8.291} |
HF Open LLM v2 | Qwen | Qwen/Qwen1.5-4B-Chat | e3417d3e-7883-45a7-a631-9e5d105788c4 | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-4B-Chat/1762652579.837912 | 1762652579.837912 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-4B-Chat | Qwen/Qwen1.5-4B-Chat | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31566576683200576}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.95} |
HF Open LLM v2 | Qwen | Qwen/Qwen2-Math-72B-Instruct | 1c7bb42e-aa1c-4522-a4b0-bcc460876125 | 0.0.1 | hfopenllm_v2/Qwen_Qwen2-Math-72B-Instruct/1762652579.841145 | 1762652579.8411462 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2-Math-72B-Instruct | Qwen/Qwen2-Math-72B-Instruct | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.569381463405985}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706} |
HF Open LLM v2 | Qwen | Qwen/Qwen1.5-MoE-A2.7B-Chat | daec0873-964e-459e-a1a1-49da96cd17cf | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-MoE-A2.7B-Chat/1762652579.838758 | 1762652579.838758 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-MoE-A2.7B-Chat | Qwen/Qwen1.5-MoE-A2.7B-Chat | Qwen | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.37953851336675576}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2MoeForCausalLM", "params_billions": 14.316} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.