_leaderboard
stringclasses
1 value
_developer
stringclasses
559 values
_model
stringlengths
9
102
_uuid
stringlengths
36
36
schema_version
stringclasses
1 value
evaluation_id
stringlengths
35
133
retrieved_timestamp
stringlengths
13
18
source_data
stringclasses
1 value
evaluation_source_name
stringclasses
1 value
evaluation_source_type
stringclasses
1 value
source_organization_name
stringclasses
1 value
source_organization_url
null
source_organization_logo_url
null
evaluator_relationship
stringclasses
1 value
model_name
stringlengths
4
102
model_id
stringlengths
9
102
model_developer
stringclasses
559 values
model_inference_platform
stringclasses
1 value
evaluation_results
stringlengths
1.35k
1.41k
additional_details
stringclasses
660 values
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Hyperionv5
5b1e2a5e-cd92-4ad4-b12d-0540461f9f5e
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Hyperionv5/1762652579.517704
1762652579.517704
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Hyperionv5
CultriX/Qwen2.5-14B-Hyperionv5
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6729211824625327}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Emerged
15af5216-fc3d-4102-bbed-eb5b7d0ecf48
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Emerged/1762652579.516177
1762652579.516178
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Emerged
CultriX/Qwen2.5-14B-Emerged
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7000237148543642}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-HyperMarck-dl
5b6ef372-86e5-4fc1-85ba-5a76517bb10f
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-HyperMarck-dl/1762652579.5170581
1762652579.517059
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-HyperMarck-dl
CultriX/Qwen2.5-14B-HyperMarck-dl
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6650276821057017}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-ReasoningMerge
df6199fa-3797-4b88-b5fc-e429f513932b
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-ReasoningMerge/1762652579.518682
1762652579.518684
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-ReasoningMerge
CultriX/Qwen2.5-14B-ReasoningMerge
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46054690443578594}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/SeQwence-14Bv2
49eccc70-6321-451b-87e9-29907cfb53a0
0.0.1
hfopenllm_v2/CultriX_SeQwence-14Bv2/1762652579.5228019
1762652579.5228028
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/SeQwence-14Bv2
CultriX/SeQwence-14Bv2
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5785992278266112}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Broca
4429613e-2db7-4061-931f-eaa70d202b71
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Broca/1762652579.5150259
1762652579.5150259
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Broca
CultriX/Qwen2.5-14B-Broca
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.560414145578177}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/SeQwence-14Bv1
f4505219-fc0d-4f7b-ad71-3c9fef064c28
0.0.1
hfopenllm_v2/CultriX_SeQwence-14Bv1/1762652579.522592
1762652579.522593
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/SeQwence-14Bv1
CultriX/SeQwence-14Bv1
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6678003253589365}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Hyper
8412921a-ad8c-4106-a3a1-9259d2ddb074
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Hyper/1762652579.516851
1762652579.516851
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Hyper
CultriX/Qwen2.5-14B-Hyper
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5391317260424563}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Ultimav2
b76ac8f6-7355-4bbf-ad8f-d8fc967120a1
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Ultimav2/1762652579.519061
1762652579.5190778
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Ultimav2
CultriX/Qwen2.5-14B-Ultimav2
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5500228283177524}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Unity
efd5d269-fc83-43f0-9054-dc3bdf40f180
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Unity/1762652579.519516
1762652579.519517
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Unity
CultriX/Qwen2.5-14B-Unity
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6738952645646883}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/SeQwence-14B-v5
6a7ae44e-93f6-4371-b3a6-585a099aa7c7
0.0.1
hfopenllm_v2/CultriX_SeQwence-14B-v5/1762652579.522369
1762652579.522369
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/SeQwence-14B-v5
CultriX/SeQwence-14B-v5
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5919881470055011}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Hyperionv4
7c4a43f8-be43-44d7-a514-f02b70ec367c
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Hyperionv4/1762652579.517484
1762652579.517484
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Hyperionv4
CultriX/Qwen2.5-14B-Hyperionv4
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5415796752616391}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Hyperionv3
d6700ad3-d858-4420-96b1-d690984ebcaa
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Hyperionv3/1762652579.517266
1762652579.517267
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Hyperionv3
CultriX/Qwen2.5-14B-Hyperionv3
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6836371937570092}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Wernicke-SLERP
8359ce66-d904-4092-92be-5e2dbb372677
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Wernicke-SLERP/1762652579.5203562
1762652579.5203571
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Wernicke-SLERP
CultriX/Qwen2.5-14B-Wernicke-SLERP
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5588904107767391}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.491}
HF Open LLM v2
alibaba
CultriX/SeQwence-14B
b9f3e9d1-e1f9-44cd-9067-c949adfbe553
0.0.1
hfopenllm_v2/CultriX_SeQwence-14B/1762652579.521544
1762652579.521545
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/SeQwence-14B
CultriX/SeQwence-14B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5351600420218354}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Brocav3
7abe4912-4e21-4774-8011-482603f7bcc0
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Brocav3/1762652579.5155342
1762652579.515535
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Brocav3
CultriX/Qwen2.5-14B-Brocav3
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6951776841004091}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/SeQwence-14B-EvolMerge
44823eb6-717b-4508-a745-7821545dd3c2
0.0.1
hfopenllm_v2/CultriX_SeQwence-14B-EvolMerge/1762652579.5218382
1762652579.5218382
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/SeQwence-14B-EvolMerge
CultriX/SeQwence-14B-EvolMerge
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5381576439403006}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-partialmergept1
852ffa19-285b-4037-ac60-63f24cafcecb
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-partialmergept1/1762652579.5208588
1762652579.52086
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-partialmergept1
CultriX/Qwen2.5-14B-partialmergept1
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.633728507028019}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwenfinity-2.5-14B
4fba9290-886e-490d-aaeb-068f8c679006
0.0.1
hfopenllm_v2/CultriX_Qwenfinity-2.5-14B/1762652579.521086
1762652579.521087
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwenfinity-2.5-14B
CultriX/Qwenfinity-2.5-14B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4813794066410457}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/SeQwence-14B-EvolMergev1
e2621a1f-af39-48fe-a56b-18e9b396a476
0.0.1
hfopenllm_v2/CultriX_SeQwence-14B-EvolMergev1/1762652579.5221288
1762652579.52213
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/SeQwence-14B-EvolMergev1
CultriX/SeQwence-14B-EvolMergev1
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5554683794554005}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Brocav7
6966d397-d336-455a-a156-c2e6430c813f
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Brocav7/1762652579.5159612
1762652579.5159621
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Brocav7
CultriX/Qwen2.5-14B-Brocav7
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6723715297632504}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-FinalMerge
36ebe0b7-51ae-4ea5-ba42-c9fd0d717259
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-FinalMerge/1762652579.516642
1762652579.516643
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-FinalMerge
CultriX/Qwen2.5-14B-FinalMerge
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48909781601705693}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/SeQwence-14Bv3
4857c00b-e4fb-417a-8b63-a5b7e9298b40
0.0.1
hfopenllm_v2/CultriX_SeQwence-14Bv3/1762652579.523057
1762652579.523058
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/SeQwence-14Bv3
CultriX/SeQwence-14Bv3
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5719047682371663}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Brocav6
63a1000f-1de8-42ef-a905-70b78bf46417
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Brocav6/1762652579.515748
1762652579.5157492
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Brocav6
CultriX/Qwen2.5-14B-Brocav6
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6995239298394925}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-Emergedv3
7b125482-fd80-4f71-b398-9421333ee736
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-Emergedv3/1762652579.516385
1762652579.516386
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-Emergedv3
CultriX/Qwen2.5-14B-Emergedv3
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6388493641316153}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
CultriX/Qwen2.5-14B-BrocaV9
782219f0-25f7-465b-9f86-5e48c9d4703e
0.0.1
hfopenllm_v2/CultriX_Qwen2.5-14B-BrocaV9/1762652579.515307
1762652579.5153081
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CultriX/Qwen2.5-14B-BrocaV9
CultriX/Qwen2.5-14B-BrocaV9
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6762933460994606}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
HPAI-BSC/Qwen2.5-Aloe-Beta-7B
a99dbb21-4f7d-4ac0-b403-2f8bf7aa92b1
0.0.1
hfopenllm_v2/HPAI-BSC_Qwen2.5-Aloe-Beta-7B/1762652579.6368651
1762652579.636866
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
HPAI-BSC/Qwen2.5-Aloe-Beta-7B
HPAI-BSC/Qwen2.5-Aloe-Beta-7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4553506917201914}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-14B-Reflective
744cef52-b155-4bb0-9411-2eb47938b5d6
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Reflective/1762652580.038453
1762652580.038454
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-14B-Reflective
braindao/DeepSeek-R1-Distill-Qwen-14B-Reflective
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4290227706928727}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-14B-ABUB-ST
46a36382-df06-4dc1-93ae-6ae61343a969
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-ABUB-ST/1762652580.036823
1762652580.036824
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-14B-ABUB-ST
braindao/DeepSeek-R1-Distill-Qwen-14B-ABUB-ST
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3751922676276723}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt
244417b6-88a2-483f-adba-c1d944c9cc29
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt/1762652580.037686
1762652580.037687
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5221456845614081}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt
407adfd5-6a1f-420a-a5de-2e37740d7025
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Blunt/1762652580.0370848
1762652580.037087
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5611632690151022}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored
41186ba2-77da-496c-afd0-c0f11ea05c9b
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored/1762652580.037415
1762652580.037416
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5421791956453321}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-7B
b4c9ec76-b126-4715-b3cf-c0d8a8a61d44
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-7B/1762652580.0386932
1762652580.038694
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-7B
braindao/DeepSeek-R1-Distill-Qwen-7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.39679938119744496}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-1.5B-Blunt
d75b9105-a60d-49d9-8606-7b23ff5d3d1a
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-1.5B-Blunt/1762652580.03596
1762652580.0359628
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-1.5B-Blunt
braindao/DeepSeek-R1-Distill-Qwen-1.5B-Blunt
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.261136008014291}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-1.5B-Reflective
40933520-61e0-4cbe-b6b2-b4d19063a1b9
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-1.5B-Reflective/1762652580.0363572
1762652580.0363579
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-1.5B-Reflective
braindao/DeepSeek-R1-Distill-Qwen-1.5B-Reflective
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30327641768285923}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-7B-Blunt
678a08d8-3089-4d97-879d-c5485344de05
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-7B-Blunt/1762652580.03893
1762652580.038931
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-7B-Blunt
braindao/DeepSeek-R1-Distill-Qwen-7B-Blunt
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4266246891581005}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt-Reflective
269f307e-3af1-47a2-92ec-00a59b4725ac
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt-Reflective/1762652580.03794
1762652580.037941
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt-Reflective
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt-Reflective
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.554044380022784}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
braindao/Qwen2.5-14B
7be8016c-2454-4228-b10d-badba12e845b
0.0.1
hfopenllm_v2/braindao_Qwen2.5-14B/1762652580.039853
1762652580.039854
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/Qwen2.5-14B
braindao/Qwen2.5-14B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.540854931581537}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-7B-ORPO-Uncensored
9c8db160-fc92-473f-a766-fb00fc099f6e
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-7B-ORPO-Uncensored/1762652580.03921
1762652580.039211
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-7B-ORPO-Uncensored
braindao/DeepSeek-R1-Distill-Qwen-7B-ORPO-Uncensored
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3654503384353515}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-7B-Reflective
fd05a73b-5b6a-460e-85d5-547710ab6bac
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-7B-Reflective/1762652580.039571
1762652580.039572
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-7B-Reflective
braindao/DeepSeek-R1-Distill-Qwen-7B-Reflective
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3921783091087204}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Reflective
1bf5eb2a-c0e2-4bfc-9ae1-ec5737974cbe
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Reflective/1762652580.038195
1762652580.038196
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Reflective
braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Reflective
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5139274901705253}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
braindao/DeepSeek-R1-Distill-Qwen-14B
f269f0cb-4f9b-4f29-84c2-a4f31ff08290
0.0.1
hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B/1762652580.036597
1762652580.036598
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
braindao/DeepSeek-R1-Distill-Qwen-14B
braindao/DeepSeek-R1-Distill-Qwen-14B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4171575863154209}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
jebish7/qwen2.5-0.5B-IHA-Hin
5849d742-02eb-4370-8c97-efc5eec4f1ed
0.0.1
hfopenllm_v2/jebish7_qwen2.5-0.5B-IHA-Hin/1762652580.28294
1762652580.28294
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jebish7/qwen2.5-0.5B-IHA-Hin
jebish7/qwen2.5-0.5B-IHA-Hin
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14163419726326149}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.494}
HF Open LLM v2
alibaba
Lambent/qwen2.5-reinstruct-alternate-lumen-14B
974e902e-0959-42d0-98f8-288e1a6ce887
0.0.1
hfopenllm_v2/Lambent_qwen2.5-reinstruct-alternate-lumen-14B/1762652579.707211
1762652579.707212
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Lambent/qwen2.5-reinstruct-alternate-lumen-14B
Lambent/qwen2.5-reinstruct-alternate-lumen-14B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47938137475232384}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
HeraiHench/DeepSeek-R1-Qwen-Coder-8B
a0730f18-1058-44b4-b6b6-0881ae2e6338
0.0.1
hfopenllm_v2/HeraiHench_DeepSeek-R1-Qwen-Coder-8B/1762652579.6392472
1762652579.639248
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
HeraiHench/DeepSeek-R1-Qwen-Coder-8B
HeraiHench/DeepSeek-R1-Qwen-Coder-8B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1869472998311148}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 8.164}
HF Open LLM v2
alibaba
HeraiHench/Marge-Qwen-Math-7B
07f4a9dc-16d7-4b75-922f-09f8e9ebed7d
0.0.1
hfopenllm_v2/HeraiHench_Marge-Qwen-Math-7B/1762652579.6397812
1762652579.639782
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
HeraiHench/Marge-Qwen-Math-7B
HeraiHench/Marge-Qwen-Math-7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.12622175826806206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
HeraiHench/Double-Down-Qwen-Math-7B
6e852e78-e666-413e-ac29-ad374bbc74f2
0.0.1
hfopenllm_v2/HeraiHench_Double-Down-Qwen-Math-7B/1762652579.63955
1762652579.639551
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
HeraiHench/Double-Down-Qwen-Math-7B
HeraiHench/Double-Down-Qwen-Math-7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1669636564316015}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
dfurman/Qwen2-72B-Orpo-v0.1
b197728d-b390-45a8-8adc-ed8567b628da
0.0.1
hfopenllm_v2/dfurman_Qwen2-72B-Orpo-v0.1/1762652580.125584
1762652580.1255848
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
dfurman/Qwen2-72B-Orpo-v0.1
dfurman/Qwen2-72B-Orpo-v0.1
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7879759039348928}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.699}
HF Open LLM v2
alibaba
CoolSpring/Qwen2-0.5B-Abyme
46d2afd2-b620-4474-ac6c-4f6bdef93d1c
0.0.1
hfopenllm_v2/CoolSpring_Qwen2-0.5B-Abyme/1762652579.5106628
1762652579.510665
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CoolSpring/Qwen2-0.5B-Abyme
CoolSpring/Qwen2-0.5B-Abyme
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.19151850423542865}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.494}
HF Open LLM v2
alibaba
CoolSpring/Qwen2-0.5B-Abyme-merge3
2a633e8b-b35a-4a26-83bb-b471bab18ed2
0.0.1
hfopenllm_v2/CoolSpring_Qwen2-0.5B-Abyme-merge3/1762652579.51142
1762652579.511421
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CoolSpring/Qwen2-0.5B-Abyme-merge3
CoolSpring/Qwen2-0.5B-Abyme-merge3
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.23860468002677343}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.63}
HF Open LLM v2
alibaba
CoolSpring/Qwen2-0.5B-Abyme-merge2
2121d736-eec6-4a86-bae0-cd032f9eb603
0.0.1
hfopenllm_v2/CoolSpring_Qwen2-0.5B-Abyme-merge2/1762652579.511093
1762652579.511094
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CoolSpring/Qwen2-0.5B-Abyme-merge2
CoolSpring/Qwen2-0.5B-Abyme-merge2
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2021846478454944}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.63}
HF Open LLM v2
alibaba
qingy2024/Qwen2.5-4B
c332cc18-e556-4b23-a45d-df26c250faa2
0.0.1
hfopenllm_v2/qingy2024_Qwen2.5-4B/1762652580.486805
1762652580.486807
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
qingy2024/Qwen2.5-4B
qingy2024/Qwen2.5-4B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.21584839337402537}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 4.168}
HF Open LLM v2
alibaba
rombodawg/Rombos-Coder-V2.5-Qwen-7b
ca077d1a-a122-4040-b7d9-924773ce67ca
0.0.1
hfopenllm_v2/rombodawg_Rombos-Coder-V2.5-Qwen-7b/1762652580.4966788
1762652580.49668
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-Coder-V2.5-Qwen-7b
rombodawg/Rombos-Coder-V2.5-Qwen-7b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6210388436016436}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
rombodawg/Rombos-LLM-V2.6-Qwen-14b
5842364a-2721-4882-90f3-97eba7c3b93a
0.0.1
hfopenllm_v2/rombodawg_Rombos-LLM-V2.6-Qwen-14b/1762652580.499588
1762652580.4995892
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-LLM-V2.6-Qwen-14b
rombodawg/Rombos-LLM-V2.6-Qwen-14b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8431550508207113}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
3f1ffcf0-10bb-46b2-ae30-3eb958e943a1
0.0.1
hfopenllm_v2/rombodawg_Rombos-LLM-V2.5.1-Qwen-3b/1762652580.498805
1762652580.498805
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2595125378440316}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.397}
HF Open LLM v2
alibaba
rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
91240596-5842-4441-b976-01ed7545bd1f
0.0.1
hfopenllm_v2/rombodawg_Rombos-LLM-V2.5.1-Qwen-3b/1762652580.499037
1762652580.499037
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
rombodawg/Rombos-LLM-V2.5.1-Qwen-3b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2566401592219755}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.397}
HF Open LLM v2
alibaba
rombodawg/Rombos-Coder-V2.5-Qwen-14b
4f7b356a-1484-458c-8bc1-2640e039ab70
0.0.1
hfopenllm_v2/rombodawg_Rombos-Coder-V2.5-Qwen-14b/1762652580.496415
1762652580.496416
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-Coder-V2.5-Qwen-14b
rombodawg/Rombos-Coder-V2.5-Qwen-14b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7047445223119102}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
rombodawg/Rombos-LLM-V2.5-Qwen-32b
07e926c9-d8bb-41da-b41e-8fddc9fb99d8
0.0.1
hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-32b/1762652580.497819
1762652580.49782
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-LLM-V2.5-Qwen-32b
rombodawg/Rombos-LLM-V2.5-Qwen-32b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6826631116548536}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764}
HF Open LLM v2
alibaba
rombodawg/Rombos-LLM-V2.5-Qwen-14b
91ec838e-699a-4c68-aa42-a9f0b3b6b0c2
0.0.1
hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-14b/1762652580.4975061
1762652580.497507
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-LLM-V2.5-Qwen-14b
rombodawg/Rombos-LLM-V2.5-Qwen-14b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5840447789642593}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
rombodawg/Rombos-LLM-V2.5-Qwen-7b
23ec1efe-a9a1-41cb-9695-4be0ceb3c199
0.0.1
hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-7b/1762652580.498573
1762652580.498574
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-LLM-V2.5-Qwen-7b
rombodawg/Rombos-LLM-V2.5-Qwen-7b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6237117514860571}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
rombodawg/Rombos-LLM-V2.5-Qwen-0.5b
796ed438-2be4-45e6-9de9-c98ddd51f3d4
0.0.1
hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-0.5b/1762652580.4969
1762652580.4969
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-LLM-V2.5-Qwen-0.5b
rombodawg/Rombos-LLM-V2.5-Qwen-0.5b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.28466690603155187}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.63}
HF Open LLM v2
alibaba
rombodawg/Rombos-LLM-V2.5-Qwen-3b
976e132a-8352-43fd-abdf-0fc4a04e9429
0.0.1
hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-3b/1762652580.498058
1762652580.498058
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-LLM-V2.5-Qwen-3b
rombodawg/Rombos-LLM-V2.5-Qwen-3b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5342358276040905}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.397}
HF Open LLM v2
alibaba
rombodawg/Rombos-LLM-V2.5-Qwen-1.5b
51f579c0-b5b4-4e01-9c19-b68fb6a21210
0.0.1
hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-1.5b/1762652580.497122
1762652580.497123
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-LLM-V2.5-Qwen-1.5b
rombodawg/Rombos-LLM-V2.5-Qwen-1.5b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3402461025634206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777}
HF Open LLM v2
alibaba
rombodawg/Rombos-LLM-V2.5-Qwen-72b
1ae05e9f-d432-4e7f-a662-4b4a118333d9
0.0.1
hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-72b/1762652580.498325
1762652580.498326
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
rombodawg/Rombos-LLM-V2.5-Qwen-72b
rombodawg/Rombos-LLM-V2.5-Qwen-72b
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.715535889218385}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706}
HF Open LLM v2
alibaba
CombinHorizon/huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES
62b4c918-b33b-40cf-888b-42b116a9e04d
0.0.1
hfopenllm_v2/CombinHorizon_huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES/1762652579.509461
1762652579.509462
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CombinHorizon/huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES
CombinHorizon/huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8175762532303177}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
CombinHorizon/Rombos-Qwen2.5-7B-Inst-BaseMerge-TIES
3171e54f-4c6f-40cf-ba6c-ef23b803ca33
0.0.1
hfopenllm_v2/CombinHorizon_Rombos-Qwen2.5-7B-Inst-BaseMerge-TIES/1762652579.508758
1762652579.508759
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CombinHorizon/Rombos-Qwen2.5-7B-Inst-BaseMerge-TIES
CombinHorizon/Rombos-Qwen2.5-7B-Inst-BaseMerge-TIES
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7564019025075688}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
CombinHorizon/huihui-ai-abliterated-Qwen2.5-32B-Inst-BaseMerge-TIES
62faed28-8f0f-4ff8-894f-b4b5b754b4cf
0.0.1
hfopenllm_v2/CombinHorizon_huihui-ai-abliterated-Qwen2.5-32B-Inst-BaseMerge-TIES/1762652579.509247
1762652579.509248
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CombinHorizon/huihui-ai-abliterated-Qwen2.5-32B-Inst-BaseMerge-TIES
CombinHorizon/huihui-ai-abliterated-Qwen2.5-32B-Inst-BaseMerge-TIES
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8206237228331937}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764}
HF Open LLM v2
alibaba
CombinHorizon/zetasepic-abliteratedV2-Qwen2.5-32B-Inst-BaseMerge-TIES
3bf71784-e6f1-405b-ad23-e74a91df7051
0.0.1
hfopenllm_v2/CombinHorizon_zetasepic-abliteratedV2-Qwen2.5-32B-Inst-BaseMerge-TIES/1762652579.509675
1762652579.509676
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CombinHorizon/zetasepic-abliteratedV2-Qwen2.5-32B-Inst-BaseMerge-TIES
CombinHorizon/zetasepic-abliteratedV2-Qwen2.5-32B-Inst-BaseMerge-TIES
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8328136012446974}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764}
HF Open LLM v2
alibaba
CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES
dcd14b21-f2fd-4c10-bf83-b6bb946f2789
0.0.1
hfopenllm_v2/CombinHorizon_Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES/1762652579.508495
1762652579.5084958
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES
CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8239958864701216}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77}
HF Open LLM v2
alibaba
AtAndDev/Qwen2.5-1.5B-continuous-learnt
1a2d8396-4ff1-4386-a76b-d4863c7736c5
0.0.1
hfopenllm_v2/AtAndDev_Qwen2.5-1.5B-continuous-learnt/1762652579.483878
1762652579.4838789
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
AtAndDev/Qwen2.5-1.5B-continuous-learnt
AtAndDev/Qwen2.5-1.5B-continuous-learnt
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45105431366551857}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544}
HF Open LLM v2
alibaba
AtAndDev/Qwen2.5-1.5B-continuous-learnt
4f7f368f-0646-4c16-80de-69d9c5e28193
0.0.1
hfopenllm_v2/AtAndDev_Qwen2.5-1.5B-continuous-learnt/1762652579.483521
1762652579.483522
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
AtAndDev/Qwen2.5-1.5B-continuous-learnt
AtAndDev/Qwen2.5-1.5B-continuous-learnt
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4605214165081982}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544}
HF Open LLM v2
alibaba
DeepMount00/Qwen2-1.5B-Ita_v6
041f6e95-b7d1-44c6-a995-0c8257e188aa
0.0.1
hfopenllm_v2/DeepMount00_Qwen2-1.5B-Ita_v6/1762652579.553008
1762652579.5530088
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DeepMount00/Qwen2-1.5B-Ita_v6
DeepMount00/Qwen2-1.5B-Ita_v6
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.29990425404593146}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.497}
HF Open LLM v2
alibaba
DeepMount00/Qwen2-1.5B-Ita_v5
04f0529b-474c-42d2-99a8-e3bdd5c18eaf
0.0.1
hfopenllm_v2/DeepMount00_Qwen2-1.5B-Ita_v5/1762652579.552789
1762652579.55279
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DeepMount00/Qwen2-1.5B-Ita_v5
DeepMount00/Qwen2-1.5B-Ita_v5
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4987400098405564}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544}
HF Open LLM v2
alibaba
DeepMount00/Qwen2-1.5B-Ita_v3
f9cac378-3bdb-4c66-8193-502773c5c5eb
0.0.1
hfopenllm_v2/DeepMount00_Qwen2-1.5B-Ita_v3/1762652579.552576
1762652579.552577
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DeepMount00/Qwen2-1.5B-Ita_v3
DeepMount00/Qwen2-1.5B-Ita_v3
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4890479483326463}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544}
HF Open LLM v2
alibaba
DeepMount00/Qwen2-1.5B-Ita_v2
78ec8596-ee15-4e94-8bc8-77c6bdffc541
0.0.1
hfopenllm_v2/DeepMount00_Qwen2-1.5B-Ita_v2/1762652579.552372
1762652579.552373
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DeepMount00/Qwen2-1.5B-Ita_v2
DeepMount00/Qwen2-1.5B-Ita_v2
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49998891829235315}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544}
HF Open LLM v2
alibaba
DeepMount00/Qwen2-1.5B-Ita
6669c8b8-91d6-4f14-8cfb-a6422352850d
0.0.1
hfopenllm_v2/DeepMount00_Qwen2-1.5B-Ita/1762652579.5521228
1762652579.5521238
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DeepMount00/Qwen2-1.5B-Ita
DeepMount00/Qwen2-1.5B-Ita
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5173495214918638}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544}
HF Open LLM v2
alibaba
hotmailuser/QwenSlerp-3B
7f53fb66-2c19-434a-acec-7cdcf9fce04d
0.0.1
hfopenllm_v2/hotmailuser_QwenSlerp-3B/1762652580.1967301
1762652580.1967309
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenSlerp-3B
hotmailuser/QwenSlerp-3B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4333690164319561}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.397}
HF Open LLM v2
alibaba
hotmailuser/QwenSlerp3-14B
7d2c1ffb-d1e7-4c88-af08-74642ddd8741
0.0.1
hfopenllm_v2/hotmailuser_QwenSlerp3-14B/1762652580.197938
1762652580.1979399
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenSlerp3-14B
hotmailuser/QwenSlerp3-14B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6632291209546226}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
hotmailuser/QwenSlerp2-14B
6732a278-0613-40fd-bdbc-88a586631279
0.0.1
hfopenllm_v2/hotmailuser_QwenSlerp2-14B/1762652580.197355
1762652580.197356
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenSlerp2-14B
hotmailuser/QwenSlerp2-14B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7036707048409332}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
hotmailuser/QwenSlerp-7B
4f8db3ee-409a-4bac-ab0a-ee3493d1e842
0.0.1
hfopenllm_v2/hotmailuser_QwenSlerp-7B/1762652580.197109
1762652580.19711
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenSlerp-7B
hotmailuser/QwenSlerp-7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4672912317096415}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
hotmailuser/QwenStock1-14B
67fd0572-cf55-412d-8ec6-0cb168d3ed08
0.0.1
hfopenllm_v2/hotmailuser_QwenStock1-14B/1762652580.1990862
1762652580.1990871
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenStock1-14B
hotmailuser/QwenStock1-14B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6693240601603745}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
hotmailuser/QwenModelStock-1.8B
661b1590-f312-447b-a494-1d37ffd93cae
0.0.1
hfopenllm_v2/hotmailuser_QwenModelStock-1.8B/1762652580.196316
1762652580.196316
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenModelStock-1.8B
hotmailuser/QwenModelStock-1.8B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3263075306852484}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777}
HF Open LLM v2
alibaba
hotmailuser/QwenStock-0.5B
72853b4d-cc12-478f-b6f4-977b8fbabfa0
0.0.1
hfopenllm_v2/hotmailuser_QwenStock-0.5B/1762652580.198598
1762652580.1985989
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenStock-0.5B
hotmailuser/QwenStock-0.5B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20490742341431845}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.63}
HF Open LLM v2
alibaba
hotmailuser/QwenStock-1.7B
25674b98-92b5-4e2d-97ab-084eabb13db2
0.0.1
hfopenllm_v2/hotmailuser_QwenStock-1.7B/1762652580.1988428
1762652580.198844
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenStock-1.7B
hotmailuser/QwenStock-1.7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.32141163224688274}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777}
HF Open LLM v2
alibaba
hotmailuser/QwenSlerp-14B
83387977-a8cd-4cdd-abc7-301006380458
0.0.1
hfopenllm_v2/hotmailuser_QwenSlerp-14B/1762652580.1965241
1762652580.196525
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenSlerp-14B
hotmailuser/QwenSlerp-14B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7024716640735471}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766}
HF Open LLM v2
alibaba
hotmailuser/Deepseek-qwen-modelstock-2B
15a4291f-4918-43a6-b242-90db88fe4a3d
0.0.1
hfopenllm_v2/hotmailuser_Deepseek-qwen-modelstock-2B/1762652580.1914759
1762652580.191477
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/Deepseek-qwen-modelstock-2B
hotmailuser/Deepseek-qwen-modelstock-2B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.21487431127186973}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777}
HF Open LLM v2
alibaba
hotmailuser/QwenSlerp2-3B
cc53c4f9-3c1b-4b21-9aac-ea22dced76c3
0.0.1
hfopenllm_v2/hotmailuser_QwenSlerp2-3B/1762652580.197566
1762652580.197566
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenSlerp2-3B
hotmailuser/QwenSlerp2-3B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4280486885907171}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.397}
HF Open LLM v2
alibaba
hotmailuser/Qwen2.5-HomerSlerp-7B
9c7dab43-b26d-4cb4-a73c-95bb1e01ffe8
0.0.1
hfopenllm_v2/hotmailuser_Qwen2.5-HomerSlerp-7B/1762652580.1961112
1762652580.1961112
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/Qwen2.5-HomerSlerp-7B
hotmailuser/Qwen2.5-HomerSlerp-7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44878145542715553}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
hotmailuser/QwenSparse-7B
96bbc2c8-bb74-408d-8625-e6bf66b63cd0
0.0.1
hfopenllm_v2/hotmailuser_QwenSparse-7B/1762652580.198252
1762652580.198254
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/QwenSparse-7B
hotmailuser/QwenSparse-7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.10858632871891026}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
adriszmar/QAIMath-Qwen2.5-7B-TIES
78544e05-7eed-465d-9199-35b25e1bebfe
0.0.1
hfopenllm_v2/adriszmar_QAIMath-Qwen2.5-7B-TIES/1762652579.9747589
1762652579.9747598
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
adriszmar/QAIMath-Qwen2.5-7B-TIES
adriszmar/QAIMath-Qwen2.5-7B-TIES
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.174632198123202}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
adriszmar/QAIMath-Qwen2.5-7B-TIES
457f0bc3-68e1-4ecb-a983-5f504b1246cd
0.0.1
hfopenllm_v2/adriszmar_QAIMath-Qwen2.5-7B-TIES/1762652579.975151
1762652579.975153
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
adriszmar/QAIMath-Qwen2.5-7B-TIES
adriszmar/QAIMath-Qwen2.5-7B-TIES
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.16853725891745014}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
alibaba
Qwen/QwQ-32B-Preview
1326f0c0-9355-47ff-813b-0729370e1487
0.0.1
hfopenllm_v2/Qwen_QwQ-32B-Preview/1762652579.834909
1762652579.83491
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/QwQ-32B-Preview
Qwen/QwQ-32B-Preview
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4035437084713006}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764}
HF Open LLM v2
alibaba
Qwen/QwQ-32B
788241ad-d975-498e-80ef-b0d04bd8db85
0.0.1
hfopenllm_v2/Qwen_QwQ-32B/1762652579.8346298
1762652579.834631
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/QwQ-32B
Qwen/QwQ-32B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.39767372793077926}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764}
HF Open LLM v2
alibaba
Qwen/Qwen1.5-7B
102378fc-7b98-4088-a6f5-3039e7b638d5
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-7B/1762652579.838115
1762652579.8381162
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-7B
Qwen/Qwen1.5-7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2684299879874289}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.721}
HF Open LLM v2
alibaba
Qwen/Qwen2.5-1.5B
9982c576-75fd-47f6-8fe9-52b56fc58d3f
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-1.5B/1762652579.8426108
1762652579.842612
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-1.5B
Qwen/Qwen2.5-1.5B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.26743041795768563}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.5}
HF Open LLM v2
alibaba
Qwen/Qwen1.5-MoE-A2.7B
c6aa0ed8-3b79-4d73-8587-762e9469f4ce
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-MoE-A2.7B/1762652579.83854
1762652579.83854
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-MoE-A2.7B
Qwen/Qwen1.5-MoE-A2.7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.265982038768246}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2MoeForCausalLM", "params_billions": 14.316}
HF Open LLM v2
alibaba
Qwen/Qwen1.5-4B
1e3f60f2-814a-4979-87bd-f5f94d5b09cc
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-4B/1762652579.837696
1762652579.837697
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-4B
Qwen/Qwen1.5-4B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24447466056729478}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.95}
HF Open LLM v2
alibaba
Qwen/Qwen1.5-32B
b8cd9221-dd4e-4f49-b03e-f11bdd5773e4
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-32B/1762652579.837265
1762652579.837266
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-32B
Qwen/Qwen1.5-32B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.329729562006587}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.512}
HF Open LLM v2
alibaba
Qwen/Qwen1.5-1.8B
7c828833-fd36-4a84-8530-d3c1769ca822
0.0.1
hfopenllm_v2/Qwen_Qwen1.5-1.8B/1762652579.835954
1762652579.835955
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen1.5-1.8B
Qwen/Qwen1.5-1.8B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2154239639711521}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.837}
HF Open LLM v2
alibaba
Qwen/Qwen2.5-Math-7B
8fddcebe-58d2-4d40-8147-f02feabc0d9c
0.0.1
hfopenllm_v2/Qwen_Qwen2.5-Math-7B/1762652579.8480499
1762652579.848052
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Qwen/Qwen2.5-Math-7B
Qwen/Qwen2.5-Math-7B
alibaba
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24599839536873275}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}