_leaderboard
stringclasses
1 value
_developer
stringclasses
559 values
_model
stringlengths
9
102
_uuid
stringlengths
36
36
schema_version
stringclasses
1 value
evaluation_id
stringlengths
35
133
retrieved_timestamp
stringlengths
13
18
source_data
stringclasses
1 value
evaluation_source_name
stringclasses
1 value
evaluation_source_type
stringclasses
1 value
source_organization_name
stringclasses
1 value
source_organization_url
null
source_organization_logo_url
null
evaluator_relationship
stringclasses
1 value
model_name
stringlengths
4
102
model_id
stringlengths
9
102
model_developer
stringclasses
559 values
model_inference_platform
stringclasses
1 value
evaluation_results
stringlengths
1.35k
1.41k
additional_details
stringclasses
660 values
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-gamma-8B
d25510e4-6549-4f64-8ec4-37ac8671050c
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-8B/1762652580.237391
1762652580.237392
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-gamma-8B
jaspionjader/Kosmos-EVAA-gamma-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45722460848326885}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/slu-6
0e1cd676-f95b-4562-8c5d-e932f148dc23
0.0.1
hfopenllm_v2/jaspionjader_slu-6/1762652580.276035
1762652580.276036
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/slu-6
jaspionjader/slu-6
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.41166216749336204}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-54
9a2d7235-84cf-43f6-8855-68d0bf85e6e3
0.0.1
hfopenllm_v2/jaspionjader_bh-54/1762652580.258788
1762652580.258792
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-54
jaspionjader/bh-54
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48405231452545916}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/kstc-1-8b
cd7e14cb-b1f1-47d8-81a9-960da8ac4e05
0.0.1
hfopenllm_v2/jaspionjader_kstc-1-8b/1762652580.2702851
1762652580.270286
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/kstc-1-8b
jaspionjader/kstc-1-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4642936297911763}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/kstc-11-8b
41b46842-dffa-4791-8225-99d676f563c9
0.0.1
hfopenllm_v2/jaspionjader_kstc-11-8b/1762652580.270522
1762652580.270522
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/kstc-11-8b
jaspionjader/kstc-11-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4757343847657549}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-7
b63d1462-f84b-4d20-86d6-1a54cf4eb81f
0.0.1
hfopenllm_v2/jaspionjader_bh-7/1762652580.261788
1762652580.261791
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-7
jaspionjader/bh-7
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4623953332712758}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-TSN-light-8B
9f0aa20f-8687-4c21-b222-39a322f90842
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-TSN-light-8B/1762652580.236298
1762652580.236299
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-TSN-light-8B
jaspionjader/Kosmos-EVAA-TSN-light-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46849027247702757}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-v2-8B
8f16aed2-8b31-48cc-b874-8d437f26f3db
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-v2-8B/1762652580.241379
1762652580.2413802
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-v2-8B
jaspionjader/Kosmos-EVAA-v2-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4395891789341171}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-gamma-v17-8B
bd4cc259-d535-437a-afc5-d74a60154b07
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-v17-8B/1762652580.239734
1762652580.239739
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-gamma-v17-8B
jaspionjader/Kosmos-EVAA-gamma-v17-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4462337708391512}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-v10-8B
c57d95da-1b6f-4ce7-8c42-f1129fc1e55e
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-v10-8B/1762652580.2406652
1762652580.2406662
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-v10-8B
jaspionjader/Kosmos-EVAA-v10-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4261503920708165}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/slu-32
1095577f-7b50-4854-9c7c-5beb59206e60
0.0.1
hfopenllm_v2/jaspionjader_slu-32/1762652580.274382
1762652580.274383
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/slu-32
jaspionjader/slu-32
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45155409868039026}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-32
a56c62cc-c318-4de4-b6c7-0fa10229a127
0.0.1
hfopenllm_v2/jaspionjader_bh-32/1762652580.253373
1762652580.2533739
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-32
jaspionjader/bh-32
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4635943740386619}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-29
012eeeed-c556-460d-82f6-34bdc31da5cf
0.0.1
hfopenllm_v2/jaspionjader_bh-29/1762652580.252519
1762652580.2525198
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-29
jaspionjader/bh-29
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46881496651107946}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bbb-3
94668ddb-d2fb-44e2-8ed7-10179d145366
0.0.1
hfopenllm_v2/jaspionjader_bbb-3/1762652580.24635
1762652580.246351
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bbb-3
jaspionjader/bbb-3
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.416832892281369}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-64
90830134-43d5-4d0c-9a93-4be2c1c7dba8
0.0.1
hfopenllm_v2/jaspionjader_bh-64/1762652580.261374
1762652580.261375
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-64
jaspionjader/bh-64
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.41401038134372353}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/slu-13
1a1eaa84-9926-4c4b-b254-96cd667c25ac
0.0.1
hfopenllm_v2/jaspionjader_slu-13/1762652580.272234
1762652580.272237
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/slu-13
jaspionjader/slu-13
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4378404854674486}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bbb-6
ef8025de-fe9f-4a79-97f6-c26c18ab049a
0.0.1
hfopenllm_v2/jaspionjader_bbb-6/1762652580.247051
1762652580.247051
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bbb-6
jaspionjader/bbb-6
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48797413086166924}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-47
9f30c4d4-4a3c-459e-8444-e143ef75f84e
0.0.1
hfopenllm_v2/jaspionjader_bh-47/1762652580.256935
1762652580.2569358
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-47
jaspionjader/bh-47
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46516797652451053}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-31
a637936e-646b-4c21-964a-61e253fd3705
0.0.1
hfopenllm_v2/jaspionjader_bh-31/1762652580.253162
1762652580.253163
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-31
jaspionjader/bh-31
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4727367828472896}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/slu-11
0091eabc-3888-4e1a-a29d-8c4e98b599f2
0.0.1
hfopenllm_v2/jaspionjader_slu-11/1762652580.272018
1762652580.272018
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/slu-11
jaspionjader/slu-11
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.372519359743259}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-41
db0c4182-7391-40e7-ad6e-5374c8eb28e1
0.0.1
hfopenllm_v2/jaspionjader_bh-41/1762652580.2558541
1762652580.2558541
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-41
jaspionjader/bh-41
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4739856912990864}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-62
85bd08bf-bdc3-42fb-b8f9-3d83e32921bc
0.0.1
hfopenllm_v2/jaspionjader_bh-62/1762652580.260948
1762652580.260949
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-62
jaspionjader/bh-62
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.41498446344587914}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-Aurora_faustus-8B
0563ee22-d981-45cb-83f8-7dbdb2734d10
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-Aurora_faustus-8B/1762652580.231864
1762652580.2318652
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-Aurora_faustus-8B
jaspionjader/Kosmos-Aurora_faustus-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.443236168920686}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/gamma-Kosmos-EVAA-8B
6e5584a8-5b8e-48ce-8b80-2d39a74a9b0d
0.0.1
hfopenllm_v2/jaspionjader_gamma-Kosmos-EVAA-8B/1762652580.268576
1762652580.268577
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/gamma-Kosmos-EVAA-8B
jaspionjader/gamma-Kosmos-EVAA-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42500121898784116}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-PRP-v34-8B
11486e0e-a9e3-43b0-b26e-299a86555d16
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v34-8B/1762652580.235871
1762652580.235871
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-PRP-v34-8B
jaspionjader/Kosmos-EVAA-PRP-v34-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45625052638111324}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/f-7-8b
e8c5d934-c9b6-460c-bd45-c4a3e2d26bed
0.0.1
hfopenllm_v2/jaspionjader_f-7-8b/1762652580.2664478
1762652580.266449
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/f-7-8b
jaspionjader/f-7-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4462337708391512}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/test-10
2bcc7f9a-9c36-487e-8522-bfbe1910b857
0.0.1
hfopenllm_v2/jaspionjader_test-10/1762652580.2777631
1762652580.277764
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/test-10
jaspionjader/test-10
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4578241288669619}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-gamma-v15-8B
24e11e0c-fb61-46c1-a05e-c533eb392195
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-v15-8B/1762652580.239064
1762652580.2390652
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-gamma-v15-8B
jaspionjader/Kosmos-EVAA-gamma-v15-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4654428028741517}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/slu-37
e64e5fe0-c726-4b9d-9d7b-952e7c7508ab
0.0.1
hfopenllm_v2/jaspionjader_slu-37/1762652580.2757561
1762652580.275757
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/slu-37
jaspionjader/slu-37
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4533526598314694}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/fct-9-8b
4d1ddf64-4626-4877-a0fa-84e06f6cf977
0.0.1
hfopenllm_v2/jaspionjader_fct-9-8b/1762652580.267691
1762652580.267692
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/fct-9-8b
jaspionjader/fct-9-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4353925362482657}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-PRP-v23-8B
5d5ae047-72d1-4083-8e28-dcce7337ed25
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v23-8B/1762652580.233495
1762652580.233495
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-PRP-v23-8B
jaspionjader/Kosmos-EVAA-PRP-v23-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4040933611705829}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/f-8-8b
dad898e1-ee18-4864-b432-462d17ac8006
0.0.1
hfopenllm_v2/jaspionjader_f-8-8b/1762652580.266931
1762652580.266932
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/f-8-8b
jaspionjader/f-8-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4739358236146758}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Auro-Kosmos-EVAA-v2.3-8B
9c7ee100-754e-4665-8527-452021a2243b
0.0.1
hfopenllm_v2/jaspionjader_Auro-Kosmos-EVAA-v2.3-8B/1762652580.231667
1762652580.231667
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Auro-Kosmos-EVAA-v2.3-8B
jaspionjader/Auro-Kosmos-EVAA-v2.3-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42712447417297217}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-61
00b1b367-c4eb-4048-b80d-a8253e7c2048
0.0.1
hfopenllm_v2/jaspionjader_bh-61/1762652580.260743
1762652580.260743
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-61
jaspionjader/bh-61
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42467652495378927}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-18
6aad7ade-7bd0-4515-b4ac-2299c58da098
0.0.1
hfopenllm_v2/jaspionjader_bh-18/1762652580.249514
1762652580.249515
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-18
jaspionjader/bh-18
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47246195649764844}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-59
974b1542-8716-4ea3-b097-f9893c9c9656
0.0.1
hfopenllm_v2/jaspionjader_bh-59/1762652580.260088
1762652580.2600892
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-59
jaspionjader/bh-59
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43414362779646887}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-PRP-v25-8B
81c8704c-7124-42d1-b320-77e31e35898b
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v25-8B/1762652580.23391
1762652580.23391
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-PRP-v25-8B
jaspionjader/Kosmos-EVAA-PRP-v25-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4420869958377106}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/knf-2-8b
267e641c-7fbd-40d3-a9b7-eb3621240b2a
0.0.1
hfopenllm_v2/jaspionjader_knf-2-8b/1762652580.269415
1762652580.2694159
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/knf-2-8b
jaspionjader/knf-2-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42500121898784116}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/gamma-Kosmos-EVAA-v3-8B
d461545f-ebcb-49e2-94ce-a6591e31a94a
0.0.1
hfopenllm_v2/jaspionjader_gamma-Kosmos-EVAA-v3-8B/1762652580.269119
1762652580.26912
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/gamma-Kosmos-EVAA-v3-8B
jaspionjader/gamma-Kosmos-EVAA-v3-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43326928106313467}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bbb-1
b6ca35e1-8680-49e8-a6dd-963214be7411
0.0.1
hfopenllm_v2/jaspionjader_bbb-1/1762652580.2456498
1762652580.245653
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bbb-1
jaspionjader/bbb-1
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4864005283758206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/knfp-2-8b
0bd6a333-afc0-43a4-9d14-fa44c2364184
0.0.1
hfopenllm_v2/jaspionjader_knfp-2-8b/1762652580.2696629
1762652580.269664
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/knfp-2-8b
jaspionjader/knfp-2-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5327120928026525}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/TSN-Kosmos-EVAA-v2-8B
2ce2b8e4-0cd4-4001-8790-ad5e26e3e45c
0.0.1
hfopenllm_v2/jaspionjader_TSN-Kosmos-EVAA-v2-8B/1762652580.2454138
1762652580.245415
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/TSN-Kosmos-EVAA-v2-8B
jaspionjader/TSN-Kosmos-EVAA-v2-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46669171132594844}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-24
0b27b829-6588-4f7b-80fe-6e6767287a38
0.0.1
hfopenllm_v2/jaspionjader_bh-24/1762652580.251392
1762652580.251392
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-24
jaspionjader/bh-24
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4715377420799035}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-26
0218b7de-bbd7-4196-8fec-3f6fb790a3c1
0.0.1
hfopenllm_v2/jaspionjader_bh-26/1762652580.251851
1762652580.251852
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-26
jaspionjader/bh-26
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4690897928607206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/knfp-3-8b
38a5c599-a098-42f4-a7cb-acee487e382a
0.0.1
hfopenllm_v2/jaspionjader_knfp-3-8b/1762652580.2700531
1762652580.2700539
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/knfp-3-8b
jaspionjader/knfp-3-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49456885508229276}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-PRP-v29-8B
86e94a19-e497-4539-802b-597ce0e0ced0
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v29-8B/1762652580.234771
1762652580.234771
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-PRP-v29-8B
jaspionjader/Kosmos-EVAA-PRP-v29-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4487315877427448}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-TSN-v21-8B
d9c819c2-a3f6-481e-bd71-47912aef9847
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-TSN-v21-8B/1762652580.2369542
1762652580.236955
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-TSN-v21-8B
jaspionjader/Kosmos-EVAA-TSN-v21-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46701640536000033}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-12
736ee66e-bd19-4275-afaf-73c2112c2fbd
0.0.1
hfopenllm_v2/jaspionjader_bh-12/1762652580.248367
1762652580.248368
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-12
jaspionjader/bh-12
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47338617091539337}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/slu-34
050afa51-be7c-4cad-ae8b-bd63384df297
0.0.1
hfopenllm_v2/jaspionjader_slu-34/1762652580.2749598
1762652580.274961
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/slu-34
jaspionjader/slu-34
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4350678422142138}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-PRP-v31-8B
0757cecd-bc5f-4095-90ee-25920ae6670c
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v31-8B/1762652580.235214
1762652580.235214
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-PRP-v31-8B
jaspionjader/Kosmos-EVAA-PRP-v31-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43986400528375824}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-33
bcab8546-ea69-4207-b69b-ab982b603e55
0.0.1
hfopenllm_v2/jaspionjader_bh-33/1762652580.25359
1762652580.253591
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-33
jaspionjader/bh-33
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4685401401614383}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-15
af3bd92d-45f5-4a48-89aa-b8c956209d5a
0.0.1
hfopenllm_v2/jaspionjader_bh-15/1762652580.248791
1762652580.2487922
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-15
jaspionjader/bh-15
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47453534399836883}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-TSN-v19-8B
91c2897a-3ae3-402b-aadf-26d0b8d746c5
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-TSN-v19-8B/1762652580.236516
1762652580.2365172
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-TSN-v19-8B
jaspionjader/Kosmos-EVAA-TSN-v19-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4563502617499346}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-16
c98928d3-0d7f-429c-927c-bf8fa432101a
0.0.1
hfopenllm_v2/jaspionjader_bh-16/1762652580.2489972
1762652580.248998
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-16
jaspionjader/bh-16
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4730614768813415}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-63
c9df2e30-5e2d-42cc-8597-dc354602350a
0.0.1
hfopenllm_v2/jaspionjader_bh-63/1762652580.261157
1762652580.261157
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-63
jaspionjader/bh-63
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43077146415954115}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-49
e574e35a-56cb-471d-b4f1-df0858f5ce66
0.0.1
hfopenllm_v2/jaspionjader_bh-49/1762652580.257362
1762652580.257366
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-49
jaspionjader/bh-49
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47246195649764844}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-Fusion-8B
ac41e588-0664-44f5-9fa9-eafd6508078b
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-Fusion-8B/1762652580.23267
1762652580.232671
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-Fusion-8B
jaspionjader/Kosmos-EVAA-Fusion-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4417623018036587}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-Fusion-8B
3030519e-f137-4091-9394-26a0779f0ad9
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-Fusion-8B/1762652580.2328691
1762652580.2328691
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-Fusion-8B
jaspionjader/Kosmos-EVAA-Fusion-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43446832183052075}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-Elusive-8b
60d775f1-47a9-45ae-9b2f-75b95c9d96cd
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-Elusive-8b/1762652580.243371
1762652580.243371
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-Elusive-8b
jaspionjader/Kosmos-Elusive-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.41688275996577967}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-48
80bbd567-b13e-4ed4-ba85-9098639a3642
0.0.1
hfopenllm_v2/jaspionjader_bh-48/1762652580.257132
1762652580.257133
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-48
jaspionjader/bh-48
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46881496651107946}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-TSN-8B
75037d12-da94-4c55-8de5-a7cef098d4b0
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-TSN-8B/1762652580.236081
1762652580.2360818
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-TSN-8B
jaspionjader/Kosmos-EVAA-TSN-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47213726246359655}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/slu-25
03c03447-1bf3-4721-8f9e-5ef041ab5d7d
0.0.1
hfopenllm_v2/jaspionjader_slu-25/1762652580.27394
1762652580.273941
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/slu-25
jaspionjader/slu-25
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4500303638789523}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-53
978d4a27-17c7-4f87-b3e5-27b00ffa4d80
0.0.1
hfopenllm_v2/jaspionjader_bh-53/1762652580.25855
1762652580.2585511
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-53
jaspionjader/bh-53
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4779573753197073}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/slu-22
c0898ca4-21a7-4d83-ad2e-1aa61bd370fa
0.0.1
hfopenllm_v2/jaspionjader_slu-22/1762652580.2733881
1762652580.273391
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/slu-22
jaspionjader/slu-22
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4321201079801593}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-PRP-v27-8B
d3dcd3f0-2f43-4b82-ba29-77a69a9b3e8f
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v27-8B/1762652580.2343428
1762652580.234344
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-PRP-v27-8B
jaspionjader/Kosmos-EVAA-PRP-v27-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4378404854674486}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/f-6-8b
2a71c7d7-8ae6-45e7-ab7f-54f7d31dd131
0.0.1
hfopenllm_v2/jaspionjader_f-6-8b/1762652580.2661529
1762652580.266155
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/f-6-8b
jaspionjader/f-6-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48460196722474147}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/test-19
ab7e0f6c-bca9-4f83-a4a0-5014c46e0512
0.0.1
hfopenllm_v2/jaspionjader_test-19/1762652580.279826
1762652580.2798269
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/test-19
jaspionjader/test-19
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44008896394898867}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/TSN-Kosmos-EVAA-8B
9819f2bd-8108-4fc5-9208-ce295d860435
0.0.1
hfopenllm_v2/jaspionjader_TSN-Kosmos-EVAA-8B/1762652580.2451851
1762652580.245186
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/TSN-Kosmos-EVAA-8B
jaspionjader/TSN-Kosmos-EVAA-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49032234471203073}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/test-20
6391f921-4de7-4e83-8bb2-8d0ef0b58d8f
0.0.1
hfopenllm_v2/jaspionjader_test-20/1762652580.2800388
1762652580.28004
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/test-20
jaspionjader/test-20
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45292823042859615}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bbb-4
828a6bd0-a205-4327-bc77-2e8a84c0b69e
0.0.1
hfopenllm_v2/jaspionjader_bbb-4/1762652580.2465842
1762652580.2465851
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bbb-4
jaspionjader/bbb-4
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47675833455232114}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-v9-8B
cd0c4096-93ee-4a04-83b0-44063770e81b
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-v9-8B/1762652580.242934
1762652580.242935
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-v9-8B
jaspionjader/Kosmos-EVAA-v9-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43686640336529303}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/kstc-6-8b
f7d63a4b-070d-4581-acce-cd356a3dea47
0.0.1
hfopenllm_v2/jaspionjader_kstc-6-8b/1762652580.2711701
1762652580.2711701
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/kstc-6-8b
jaspionjader/kstc-6-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49439376410147295}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-25
a0c16d3d-e3f2-4c50-975a-70b69824b3d5
0.0.1
hfopenllm_v2/jaspionjader_bh-25/1762652580.251633
1762652580.251633
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-25
jaspionjader/bh-25
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47518473206647255}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-gamma-alt-8B
58e279d4-da0f-4e2c-a74d-c51caeaad884
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-alt-8B/1762652580.23761
1762652580.23761
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-gamma-alt-8B
jaspionjader/Kosmos-EVAA-gamma-alt-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4542270065648036}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-21
a369ff4f-7fe9-4764-be74-83563dbaf635
0.0.1
hfopenllm_v2/jaspionjader_bh-21/1762652580.25052
1762652580.2505212
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-21
jaspionjader/bh-21
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47001400727846554}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-gamma-light-alt-8B
abebffbf-48b5-4452-8c7a-bb1175a7e979
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-light-alt-8B/1762652580.238084
1762652580.238085
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-gamma-light-alt-8B
jaspionjader/Kosmos-EVAA-gamma-light-alt-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44535942410581697}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
Qwen
Qwen/Qwen1.5-0.5B-Chat
96baee1a-7ea7-454f-ac8b-fe1bead3cd93
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-0.5B-Chat/1762652579.835679
1762652579.83568
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-0.5B-Chat
Qwen/Qwen1.5-0.5B-Chat
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.18072713732895385}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.62}
HF Open LLM v2
Qwen
Qwen/Qwen2.5-0.5B-Instruct
883755e2-69eb-459b-ae7f-5548914aa65e
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-0.5B-Instruct/1762652579.842189
1762652579.84219
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-0.5B-Instruct
Qwen/Qwen2.5-0.5B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.307122878407071}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.5}
HF Open LLM v2
Qwen
Qwen/Qwen2.5-0.5B-Instruct
14d1ea99-ae05-42cd-9f2f-de1a98d9846d
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-0.5B-Instruct/1762652579.842413
1762652579.8424141
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-0.5B-Instruct
Qwen/Qwen2.5-0.5B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31529120511354314}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.494}
HF Open LLM v2
Qwen
Qwen/Qwen2.5-72B-Instruct
9ed2a831-aa5a-4e81-b8b5-397bc8b55835
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-72B-Instruct/1762652579.844789
1762652579.844789
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-72B-Instruct
Qwen/Qwen2.5-72B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.863837949972739}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706}
HF Open LLM v2
Qwen
Qwen/Qwen1.5-14B-Chat
e2cdcc99-a1b6-43ee-9cda-2e7ccbd0ad8d
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-14B-Chat/1762652579.837058
1762652579.837059
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-14B-Chat
Qwen/Qwen1.5-14B-Chat
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47680820223673187}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.167}
HF Open LLM v2
Qwen
Qwen/Qwen1.5-1.8B-Chat
d6107bde-875e-40f6-8471-3a3507758910
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-1.8B-Chat/1762652579.836214
1762652579.836215
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-1.8B-Chat
Qwen/Qwen1.5-1.8B-Chat
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20190982149585324}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.837}
HF Open LLM v2
Qwen
Qwen/Qwen2.5-32B-Instruct
c921186d-6e97-46d6-b968-894159271620
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-32B-Instruct/1762652579.843922
1762652579.843922
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-32B-Instruct
Qwen/Qwen2.5-32B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8346121623957765}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764}
HF Open LLM v2
Qwen
Qwen/Qwen2-1.5B-Instruct
984029c7-f957-4555-8460-dfecd99f44a1
0.0.1
hfopenllm_v2/Qwen_Qwen2-1.5B-Instruct/1762652579.839607
1762652579.839607
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2-1.5B-Instruct
Qwen/Qwen2-1.5B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3371232773485463}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544}
HF Open LLM v2
Qwen
Qwen/Qwen2.5-Coder-7B-Instruct
81749833-4f2a-4883-a789-c465c11b33b6
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-Coder-7B-Instruct/1762652579.8473449
1762652579.8473458
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-Coder-7B-Instruct
Qwen/Qwen2.5-Coder-7B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6147189457306613}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
Qwen
Qwen/Qwen2.5-Coder-7B-Instruct
7629f304-5235-485b-a7f6-f5a7f91fd35c
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-Coder-7B-Instruct/1762652579.847122
1762652579.847123
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-Coder-7B-Instruct
Qwen/Qwen2.5-Coder-7B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6101477413263474}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
Qwen
Qwen/Qwen2.5-7B-Instruct
7a336f2b-3b33-4fde-bce6-2d1e884a1b26
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-7B-Instruct/1762652579.845207
1762652579.8452082
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-7B-Instruct
Qwen/Qwen2.5-7B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7585251576926999}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
Qwen
Qwen/Qwen2.5-Math-72B-Instruct
64574dc3-4982-49c3-8526-09ebd5781175
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-Math-72B-Instruct/1762652579.847774
1762652579.847775
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-Math-72B-Instruct
Qwen/Qwen2.5-Math-72B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4003466358151926}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706}
HF Open LLM v2
Qwen
Qwen/Qwen2-57B-A14B-Instruct
50496313-dc6c-4456-8a8c-15cd8ddbb480
0.0.1
hfopenllm_v2/Qwen_Qwen2-57B-A14B-Instruct/1762652579.84003
1762652579.840031
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2-57B-A14B-Instruct
Qwen/Qwen2-57B-A14B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6337783747124297}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2MoeForCausalLM", "params_billions": 57.409}
HF Open LLM v2
Qwen
Qwen/Qwen2.5-14B-Instruct
1f3e04ab-9f97-4eda-9d40-669eda073ac3
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-14B-Instruct/1762652579.843263
1762652579.843264
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-14B-Instruct
Qwen/Qwen2.5-14B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8157776920792386}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
Qwen
Qwen/Qwen1.5-7B-Chat
42e3c9e4-bf1a-43ae-87e7-056f735abe03
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-7B-Chat/1762652579.838321
1762652579.838322
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-7B-Chat
Qwen/Qwen1.5-7B-Chat
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43711574178734647}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.721}
HF Open LLM v2
Qwen
Qwen/Qwen2-VL-72B-Instruct
2f749e28-b845-45ab-a628-8f9b6a9029d9
0.0.1
hfopenllm_v2/Qwen_Qwen2-VL-72B-Instruct/1762652579.841569
1762652579.8415701
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2-VL-72B-Instruct
Qwen/Qwen2-VL-72B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5982326892644849}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2VLForConditionalGeneration", "params_billions": 73.406}
HF Open LLM v2
Qwen
Qwen/Qwen1.5-32B-Chat
c14a0d32-1d27-4596-90d4-10a793aef9a2
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-32B-Chat/1762652579.8374798
1762652579.8374798
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-32B-Chat
Qwen/Qwen1.5-32B-Chat
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5532199009738605}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.512}
HF Open LLM v2
Qwen
Qwen/Qwen2.5-1.5B-Instruct
9744dd76-a8cd-4400-92a7-f10b375710ae
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-1.5B-Instruct/1762652579.842835
1762652579.842836
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-1.5B-Instruct
Qwen/Qwen2.5-1.5B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4475569267321817}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.5}
HF Open LLM v2
Qwen
Qwen/Qwen2-72B-Instruct
d9ae7c35-ac71-4703-9cfe-bf5fb5aa688e
0.0.1
hfopenllm_v2/Qwen_Qwen2-72B-Instruct/1762652579.840446
1762652579.840447
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2-72B-Instruct
Qwen/Qwen2-72B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7989168738945996}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706}
HF Open LLM v2
Qwen
Qwen/Qwen1.5-110B-Chat
7cfcae3d-b623-4cf0-9ac8-529db46d05e6
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-110B-Chat/1762652579.836649
1762652579.836649
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-110B-Chat
Qwen/Qwen1.5-110B-Chat
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5938864435254014}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 111.21}
HF Open LLM v2
Qwen
Qwen/Qwen2-VL-7B-Instruct
6dd0eebe-ef61-431d-bf7c-c170475bed5f
0.0.1
hfopenllm_v2/Qwen_Qwen2-VL-7B-Instruct/1762652579.841773
1762652579.841774
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2-VL-7B-Instruct
Qwen/Qwen2-VL-7B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4599218961245052}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2VLForConditionalGeneration", "params_billions": 8.291}
HF Open LLM v2
Qwen
Qwen/Qwen1.5-4B-Chat
e3417d3e-7883-45a7-a631-9e5d105788c4
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-4B-Chat/1762652579.837912
1762652579.837912
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-4B-Chat
Qwen/Qwen1.5-4B-Chat
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31566576683200576}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.95}
HF Open LLM v2
Qwen
Qwen/Qwen2-Math-72B-Instruct
1c7bb42e-aa1c-4522-a4b0-bcc460876125
0.0.1
hfopenllm_v2/Qwen_Qwen2-Math-72B-Instruct/1762652579.841145
1762652579.8411462
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2-Math-72B-Instruct
Qwen/Qwen2-Math-72B-Instruct
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.569381463405985}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706}
HF Open LLM v2
Qwen
Qwen/Qwen1.5-MoE-A2.7B-Chat
daec0873-964e-459e-a1a1-49da96cd17cf
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-MoE-A2.7B-Chat/1762652579.838758
1762652579.838758
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-MoE-A2.7B-Chat
Qwen/Qwen1.5-MoE-A2.7B-Chat
Qwen
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.37953851336675576}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2MoeForCausalLM", "params_billions": 14.316}