_leaderboard stringclasses 1
value | _developer stringclasses 559
values | _model stringlengths 9 102 | _uuid stringlengths 36 36 | schema_version stringclasses 1
value | evaluation_id stringlengths 35 133 | retrieved_timestamp stringlengths 13 18 | source_data stringclasses 1
value | evaluation_source_name stringclasses 1
value | evaluation_source_type stringclasses 1
value | source_organization_name stringclasses 1
value | source_organization_url null | source_organization_logo_url null | evaluator_relationship stringclasses 1
value | model_name stringlengths 4 102 | model_id stringlengths 9 102 | model_developer stringclasses 559
values | model_inference_platform stringclasses 1
value | evaluation_results stringlengths 1.35k 1.41k | additional_details stringclasses 660
values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Hyperionv5 | 5b1e2a5e-cd92-4ad4-b12d-0540461f9f5e | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Hyperionv5/1762652579.517704 | 1762652579.517704 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Hyperionv5 | CultriX/Qwen2.5-14B-Hyperionv5 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6729211824625327}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Emerged | 15af5216-fc3d-4102-bbed-eb5b7d0ecf48 | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Emerged/1762652579.516177 | 1762652579.516178 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Emerged | CultriX/Qwen2.5-14B-Emerged | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7000237148543642}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-HyperMarck-dl | 5b6ef372-86e5-4fc1-85ba-5a76517bb10f | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-HyperMarck-dl/1762652579.5170581 | 1762652579.517059 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-HyperMarck-dl | CultriX/Qwen2.5-14B-HyperMarck-dl | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6650276821057017}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-ReasoningMerge | df6199fa-3797-4b88-b5fc-e429f513932b | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-ReasoningMerge/1762652579.518682 | 1762652579.518684 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-ReasoningMerge | CultriX/Qwen2.5-14B-ReasoningMerge | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46054690443578594}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/SeQwence-14Bv2 | 49eccc70-6321-451b-87e9-29907cfb53a0 | 0.0.1 | hfopenllm_v2/CultriX_SeQwence-14Bv2/1762652579.5228019 | 1762652579.5228028 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/SeQwence-14Bv2 | CultriX/SeQwence-14Bv2 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5785992278266112}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Broca | 4429613e-2db7-4061-931f-eaa70d202b71 | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Broca/1762652579.5150259 | 1762652579.5150259 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Broca | CultriX/Qwen2.5-14B-Broca | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.560414145578177}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/SeQwence-14Bv1 | f4505219-fc0d-4f7b-ad71-3c9fef064c28 | 0.0.1 | hfopenllm_v2/CultriX_SeQwence-14Bv1/1762652579.522592 | 1762652579.522593 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/SeQwence-14Bv1 | CultriX/SeQwence-14Bv1 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6678003253589365}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Hyper | 8412921a-ad8c-4106-a3a1-9259d2ddb074 | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Hyper/1762652579.516851 | 1762652579.516851 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Hyper | CultriX/Qwen2.5-14B-Hyper | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5391317260424563}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Ultimav2 | b76ac8f6-7355-4bbf-ad8f-d8fc967120a1 | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Ultimav2/1762652579.519061 | 1762652579.5190778 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Ultimav2 | CultriX/Qwen2.5-14B-Ultimav2 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5500228283177524}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Unity | efd5d269-fc83-43f0-9054-dc3bdf40f180 | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Unity/1762652579.519516 | 1762652579.519517 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Unity | CultriX/Qwen2.5-14B-Unity | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6738952645646883}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/SeQwence-14B-v5 | 6a7ae44e-93f6-4371-b3a6-585a099aa7c7 | 0.0.1 | hfopenllm_v2/CultriX_SeQwence-14B-v5/1762652579.522369 | 1762652579.522369 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/SeQwence-14B-v5 | CultriX/SeQwence-14B-v5 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5919881470055011}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Hyperionv4 | 7c4a43f8-be43-44d7-a514-f02b70ec367c | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Hyperionv4/1762652579.517484 | 1762652579.517484 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Hyperionv4 | CultriX/Qwen2.5-14B-Hyperionv4 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5415796752616391}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Hyperionv3 | d6700ad3-d858-4420-96b1-d690984ebcaa | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Hyperionv3/1762652579.517266 | 1762652579.517267 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Hyperionv3 | CultriX/Qwen2.5-14B-Hyperionv3 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6836371937570092}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Wernicke-SLERP | 8359ce66-d904-4092-92be-5e2dbb372677 | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Wernicke-SLERP/1762652579.5203562 | 1762652579.5203571 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Wernicke-SLERP | CultriX/Qwen2.5-14B-Wernicke-SLERP | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5588904107767391}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.491} |
HF Open LLM v2 | alibaba | CultriX/SeQwence-14B | b9f3e9d1-e1f9-44cd-9067-c949adfbe553 | 0.0.1 | hfopenllm_v2/CultriX_SeQwence-14B/1762652579.521544 | 1762652579.521545 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/SeQwence-14B | CultriX/SeQwence-14B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5351600420218354}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Brocav3 | 7abe4912-4e21-4774-8011-482603f7bcc0 | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Brocav3/1762652579.5155342 | 1762652579.515535 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Brocav3 | CultriX/Qwen2.5-14B-Brocav3 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6951776841004091}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/SeQwence-14B-EvolMerge | 44823eb6-717b-4508-a745-7821545dd3c2 | 0.0.1 | hfopenllm_v2/CultriX_SeQwence-14B-EvolMerge/1762652579.5218382 | 1762652579.5218382 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/SeQwence-14B-EvolMerge | CultriX/SeQwence-14B-EvolMerge | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5381576439403006}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-partialmergept1 | 852ffa19-285b-4037-ac60-63f24cafcecb | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-partialmergept1/1762652579.5208588 | 1762652579.52086 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-partialmergept1 | CultriX/Qwen2.5-14B-partialmergept1 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.633728507028019}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwenfinity-2.5-14B | 4fba9290-886e-490d-aaeb-068f8c679006 | 0.0.1 | hfopenllm_v2/CultriX_Qwenfinity-2.5-14B/1762652579.521086 | 1762652579.521087 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwenfinity-2.5-14B | CultriX/Qwenfinity-2.5-14B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4813794066410457}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/SeQwence-14B-EvolMergev1 | e2621a1f-af39-48fe-a56b-18e9b396a476 | 0.0.1 | hfopenllm_v2/CultriX_SeQwence-14B-EvolMergev1/1762652579.5221288 | 1762652579.52213 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/SeQwence-14B-EvolMergev1 | CultriX/SeQwence-14B-EvolMergev1 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5554683794554005}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Brocav7 | 6966d397-d336-455a-a156-c2e6430c813f | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Brocav7/1762652579.5159612 | 1762652579.5159621 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Brocav7 | CultriX/Qwen2.5-14B-Brocav7 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6723715297632504}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-FinalMerge | 36ebe0b7-51ae-4ea5-ba42-c9fd0d717259 | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-FinalMerge/1762652579.516642 | 1762652579.516643 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-FinalMerge | CultriX/Qwen2.5-14B-FinalMerge | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48909781601705693}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/SeQwence-14Bv3 | 4857c00b-e4fb-417a-8b63-a5b7e9298b40 | 0.0.1 | hfopenllm_v2/CultriX_SeQwence-14Bv3/1762652579.523057 | 1762652579.523058 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/SeQwence-14Bv3 | CultriX/SeQwence-14Bv3 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5719047682371663}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Brocav6 | 63a1000f-1de8-42ef-a905-70b78bf46417 | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Brocav6/1762652579.515748 | 1762652579.5157492 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Brocav6 | CultriX/Qwen2.5-14B-Brocav6 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6995239298394925}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-Emergedv3 | 7b125482-fd80-4f71-b398-9421333ee736 | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-Emergedv3/1762652579.516385 | 1762652579.516386 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-Emergedv3 | CultriX/Qwen2.5-14B-Emergedv3 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6388493641316153}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | CultriX/Qwen2.5-14B-BrocaV9 | 782219f0-25f7-465b-9f86-5e48c9d4703e | 0.0.1 | hfopenllm_v2/CultriX_Qwen2.5-14B-BrocaV9/1762652579.515307 | 1762652579.5153081 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CultriX/Qwen2.5-14B-BrocaV9 | CultriX/Qwen2.5-14B-BrocaV9 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6762933460994606}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | HPAI-BSC/Qwen2.5-Aloe-Beta-7B | a99dbb21-4f7d-4ac0-b403-2f8bf7aa92b1 | 0.0.1 | hfopenllm_v2/HPAI-BSC_Qwen2.5-Aloe-Beta-7B/1762652579.6368651 | 1762652579.636866 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | HPAI-BSC/Qwen2.5-Aloe-Beta-7B | HPAI-BSC/Qwen2.5-Aloe-Beta-7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4553506917201914}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-14B-Reflective | 744cef52-b155-4bb0-9411-2eb47938b5d6 | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Reflective/1762652580.038453 | 1762652580.038454 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-14B-Reflective | braindao/DeepSeek-R1-Distill-Qwen-14B-Reflective | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4290227706928727}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-14B-ABUB-ST | 46a36382-df06-4dc1-93ae-6ae61343a969 | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-ABUB-ST/1762652580.036823 | 1762652580.036824 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-14B-ABUB-ST | braindao/DeepSeek-R1-Distill-Qwen-14B-ABUB-ST | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3751922676276723}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt | 244417b6-88a2-483f-adba-c1d944c9cc29 | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt/1762652580.037686 | 1762652580.037687 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5221456845614081}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt | 407adfd5-6a1f-420a-a5de-2e37740d7025 | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Blunt/1762652580.0370848 | 1762652580.037087 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5611632690151022}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored | 41186ba2-77da-496c-afd0-c0f11ea05c9b | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored/1762652580.037415 | 1762652580.037416 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5421791956453321}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-7B | b4c9ec76-b126-4715-b3cf-c0d8a8a61d44 | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-7B/1762652580.0386932 | 1762652580.038694 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-7B | braindao/DeepSeek-R1-Distill-Qwen-7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.39679938119744496}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-1.5B-Blunt | d75b9105-a60d-49d9-8606-7b23ff5d3d1a | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-1.5B-Blunt/1762652580.03596 | 1762652580.0359628 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-1.5B-Blunt | braindao/DeepSeek-R1-Distill-Qwen-1.5B-Blunt | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.261136008014291}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-1.5B-Reflective | 40933520-61e0-4cbe-b6b2-b4d19063a1b9 | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-1.5B-Reflective/1762652580.0363572 | 1762652580.0363579 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-1.5B-Reflective | braindao/DeepSeek-R1-Distill-Qwen-1.5B-Reflective | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30327641768285923}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-7B-Blunt | 678a08d8-3089-4d97-879d-c5485344de05 | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-7B-Blunt/1762652580.03893 | 1762652580.038931 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-7B-Blunt | braindao/DeepSeek-R1-Distill-Qwen-7B-Blunt | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4266246891581005}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt-Reflective | 269f307e-3af1-47a2-92ec-00a59b4725ac | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt-Reflective/1762652580.03794 | 1762652580.037941 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt-Reflective | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Blunt-Reflective | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.554044380022784}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | braindao/Qwen2.5-14B | 7be8016c-2454-4228-b10d-badba12e845b | 0.0.1 | hfopenllm_v2/braindao_Qwen2.5-14B/1762652580.039853 | 1762652580.039854 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/Qwen2.5-14B | braindao/Qwen2.5-14B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.540854931581537}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-7B-ORPO-Uncensored | 9c8db160-fc92-473f-a766-fb00fc099f6e | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-7B-ORPO-Uncensored/1762652580.03921 | 1762652580.039211 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-7B-ORPO-Uncensored | braindao/DeepSeek-R1-Distill-Qwen-7B-ORPO-Uncensored | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3654503384353515}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-7B-Reflective | fd05a73b-5b6a-460e-85d5-547710ab6bac | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-7B-Reflective/1762652580.039571 | 1762652580.039572 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-7B-Reflective | braindao/DeepSeek-R1-Distill-Qwen-7B-Reflective | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3921783091087204}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Reflective | 1bf5eb2a-c0e2-4bfc-9ae1-ec5737974cbe | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Reflective/1762652580.038195 | 1762652580.038196 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Reflective | braindao/DeepSeek-R1-Distill-Qwen-14B-Blunt-Uncensored-Reflective | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5139274901705253}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | braindao/DeepSeek-R1-Distill-Qwen-14B | f269f0cb-4f9b-4f29-84c2-a4f31ff08290 | 0.0.1 | hfopenllm_v2/braindao_DeepSeek-R1-Distill-Qwen-14B/1762652580.036597 | 1762652580.036598 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | braindao/DeepSeek-R1-Distill-Qwen-14B | braindao/DeepSeek-R1-Distill-Qwen-14B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4171575863154209}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | jebish7/qwen2.5-0.5B-IHA-Hin | 5849d742-02eb-4370-8c97-efc5eec4f1ed | 0.0.1 | hfopenllm_v2/jebish7_qwen2.5-0.5B-IHA-Hin/1762652580.28294 | 1762652580.28294 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jebish7/qwen2.5-0.5B-IHA-Hin | jebish7/qwen2.5-0.5B-IHA-Hin | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14163419726326149}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.494} |
HF Open LLM v2 | alibaba | Lambent/qwen2.5-reinstruct-alternate-lumen-14B | 974e902e-0959-42d0-98f8-288e1a6ce887 | 0.0.1 | hfopenllm_v2/Lambent_qwen2.5-reinstruct-alternate-lumen-14B/1762652579.707211 | 1762652579.707212 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Lambent/qwen2.5-reinstruct-alternate-lumen-14B | Lambent/qwen2.5-reinstruct-alternate-lumen-14B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47938137475232384}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | HeraiHench/DeepSeek-R1-Qwen-Coder-8B | a0730f18-1058-44b4-b6b6-0881ae2e6338 | 0.0.1 | hfopenllm_v2/HeraiHench_DeepSeek-R1-Qwen-Coder-8B/1762652579.6392472 | 1762652579.639248 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | HeraiHench/DeepSeek-R1-Qwen-Coder-8B | HeraiHench/DeepSeek-R1-Qwen-Coder-8B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1869472998311148}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 8.164} |
HF Open LLM v2 | alibaba | HeraiHench/Marge-Qwen-Math-7B | 07f4a9dc-16d7-4b75-922f-09f8e9ebed7d | 0.0.1 | hfopenllm_v2/HeraiHench_Marge-Qwen-Math-7B/1762652579.6397812 | 1762652579.639782 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | HeraiHench/Marge-Qwen-Math-7B | HeraiHench/Marge-Qwen-Math-7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.12622175826806206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | HeraiHench/Double-Down-Qwen-Math-7B | 6e852e78-e666-413e-ac29-ad374bbc74f2 | 0.0.1 | hfopenllm_v2/HeraiHench_Double-Down-Qwen-Math-7B/1762652579.63955 | 1762652579.639551 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | HeraiHench/Double-Down-Qwen-Math-7B | HeraiHench/Double-Down-Qwen-Math-7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1669636564316015}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | dfurman/Qwen2-72B-Orpo-v0.1 | b197728d-b390-45a8-8adc-ed8567b628da | 0.0.1 | hfopenllm_v2/dfurman_Qwen2-72B-Orpo-v0.1/1762652580.125584 | 1762652580.1255848 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | dfurman/Qwen2-72B-Orpo-v0.1 | dfurman/Qwen2-72B-Orpo-v0.1 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7879759039348928}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.699} |
HF Open LLM v2 | alibaba | CoolSpring/Qwen2-0.5B-Abyme | 46d2afd2-b620-4474-ac6c-4f6bdef93d1c | 0.0.1 | hfopenllm_v2/CoolSpring_Qwen2-0.5B-Abyme/1762652579.5106628 | 1762652579.510665 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CoolSpring/Qwen2-0.5B-Abyme | CoolSpring/Qwen2-0.5B-Abyme | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.19151850423542865}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.494} |
HF Open LLM v2 | alibaba | CoolSpring/Qwen2-0.5B-Abyme-merge3 | 2a633e8b-b35a-4a26-83bb-b471bab18ed2 | 0.0.1 | hfopenllm_v2/CoolSpring_Qwen2-0.5B-Abyme-merge3/1762652579.51142 | 1762652579.511421 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CoolSpring/Qwen2-0.5B-Abyme-merge3 | CoolSpring/Qwen2-0.5B-Abyme-merge3 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.23860468002677343}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.63} |
HF Open LLM v2 | alibaba | CoolSpring/Qwen2-0.5B-Abyme-merge2 | 2121d736-eec6-4a86-bae0-cd032f9eb603 | 0.0.1 | hfopenllm_v2/CoolSpring_Qwen2-0.5B-Abyme-merge2/1762652579.511093 | 1762652579.511094 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CoolSpring/Qwen2-0.5B-Abyme-merge2 | CoolSpring/Qwen2-0.5B-Abyme-merge2 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2021846478454944}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.63} |
HF Open LLM v2 | alibaba | qingy2024/Qwen2.5-4B | c332cc18-e556-4b23-a45d-df26c250faa2 | 0.0.1 | hfopenllm_v2/qingy2024_Qwen2.5-4B/1762652580.486805 | 1762652580.486807 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | qingy2024/Qwen2.5-4B | qingy2024/Qwen2.5-4B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.21584839337402537}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 4.168} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-Coder-V2.5-Qwen-7b | ca077d1a-a122-4040-b7d9-924773ce67ca | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-Coder-V2.5-Qwen-7b/1762652580.4966788 | 1762652580.49668 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-Coder-V2.5-Qwen-7b | rombodawg/Rombos-Coder-V2.5-Qwen-7b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6210388436016436}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-LLM-V2.6-Qwen-14b | 5842364a-2721-4882-90f3-97eba7c3b93a | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-LLM-V2.6-Qwen-14b/1762652580.499588 | 1762652580.4995892 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-LLM-V2.6-Qwen-14b | rombodawg/Rombos-LLM-V2.6-Qwen-14b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8431550508207113}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-LLM-V2.5.1-Qwen-3b | 3f1ffcf0-10bb-46b2-ae30-3eb958e943a1 | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-LLM-V2.5.1-Qwen-3b/1762652580.498805 | 1762652580.498805 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-LLM-V2.5.1-Qwen-3b | rombodawg/Rombos-LLM-V2.5.1-Qwen-3b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2595125378440316}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.397} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-LLM-V2.5.1-Qwen-3b | 91240596-5842-4441-b976-01ed7545bd1f | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-LLM-V2.5.1-Qwen-3b/1762652580.499037 | 1762652580.499037 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-LLM-V2.5.1-Qwen-3b | rombodawg/Rombos-LLM-V2.5.1-Qwen-3b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2566401592219755}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.397} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-Coder-V2.5-Qwen-14b | 4f7b356a-1484-458c-8bc1-2640e039ab70 | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-Coder-V2.5-Qwen-14b/1762652580.496415 | 1762652580.496416 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-Coder-V2.5-Qwen-14b | rombodawg/Rombos-Coder-V2.5-Qwen-14b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7047445223119102}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-LLM-V2.5-Qwen-32b | 07e926c9-d8bb-41da-b41e-8fddc9fb99d8 | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-32b/1762652580.497819 | 1762652580.49782 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-LLM-V2.5-Qwen-32b | rombodawg/Rombos-LLM-V2.5-Qwen-32b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6826631116548536}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-LLM-V2.5-Qwen-14b | 91ec838e-699a-4c68-aa42-a9f0b3b6b0c2 | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-14b/1762652580.4975061 | 1762652580.497507 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-LLM-V2.5-Qwen-14b | rombodawg/Rombos-LLM-V2.5-Qwen-14b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5840447789642593}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-LLM-V2.5-Qwen-7b | 23ec1efe-a9a1-41cb-9695-4be0ceb3c199 | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-7b/1762652580.498573 | 1762652580.498574 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-LLM-V2.5-Qwen-7b | rombodawg/Rombos-LLM-V2.5-Qwen-7b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6237117514860571}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-LLM-V2.5-Qwen-0.5b | 796ed438-2be4-45e6-9de9-c98ddd51f3d4 | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-0.5b/1762652580.4969 | 1762652580.4969 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-LLM-V2.5-Qwen-0.5b | rombodawg/Rombos-LLM-V2.5-Qwen-0.5b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.28466690603155187}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.63} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-LLM-V2.5-Qwen-3b | 976e132a-8352-43fd-abdf-0fc4a04e9429 | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-3b/1762652580.498058 | 1762652580.498058 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-LLM-V2.5-Qwen-3b | rombodawg/Rombos-LLM-V2.5-Qwen-3b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5342358276040905}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.397} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-LLM-V2.5-Qwen-1.5b | 51f579c0-b5b4-4e01-9c19-b68fb6a21210 | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-1.5b/1762652580.497122 | 1762652580.497123 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-LLM-V2.5-Qwen-1.5b | rombodawg/Rombos-LLM-V2.5-Qwen-1.5b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3402461025634206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777} |
HF Open LLM v2 | alibaba | rombodawg/Rombos-LLM-V2.5-Qwen-72b | 1ae05e9f-d432-4e7f-a662-4b4a118333d9 | 0.0.1 | hfopenllm_v2/rombodawg_Rombos-LLM-V2.5-Qwen-72b/1762652580.498325 | 1762652580.498326 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | rombodawg/Rombos-LLM-V2.5-Qwen-72b | rombodawg/Rombos-LLM-V2.5-Qwen-72b | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.715535889218385}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706} |
HF Open LLM v2 | alibaba | CombinHorizon/huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES | 62b4c918-b33b-40cf-888b-42b116a9e04d | 0.0.1 | hfopenllm_v2/CombinHorizon_huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES/1762652579.509461 | 1762652579.509462 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CombinHorizon/huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES | CombinHorizon/huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8175762532303177}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | CombinHorizon/Rombos-Qwen2.5-7B-Inst-BaseMerge-TIES | 3171e54f-4c6f-40cf-ba6c-ef23b803ca33 | 0.0.1 | hfopenllm_v2/CombinHorizon_Rombos-Qwen2.5-7B-Inst-BaseMerge-TIES/1762652579.508758 | 1762652579.508759 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CombinHorizon/Rombos-Qwen2.5-7B-Inst-BaseMerge-TIES | CombinHorizon/Rombos-Qwen2.5-7B-Inst-BaseMerge-TIES | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7564019025075688}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | CombinHorizon/huihui-ai-abliterated-Qwen2.5-32B-Inst-BaseMerge-TIES | 62faed28-8f0f-4ff8-894f-b4b5b754b4cf | 0.0.1 | hfopenllm_v2/CombinHorizon_huihui-ai-abliterated-Qwen2.5-32B-Inst-BaseMerge-TIES/1762652579.509247 | 1762652579.509248 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CombinHorizon/huihui-ai-abliterated-Qwen2.5-32B-Inst-BaseMerge-TIES | CombinHorizon/huihui-ai-abliterated-Qwen2.5-32B-Inst-BaseMerge-TIES | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8206237228331937}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764} |
HF Open LLM v2 | alibaba | CombinHorizon/zetasepic-abliteratedV2-Qwen2.5-32B-Inst-BaseMerge-TIES | 3bf71784-e6f1-405b-ad23-e74a91df7051 | 0.0.1 | hfopenllm_v2/CombinHorizon_zetasepic-abliteratedV2-Qwen2.5-32B-Inst-BaseMerge-TIES/1762652579.509675 | 1762652579.509676 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CombinHorizon/zetasepic-abliteratedV2-Qwen2.5-32B-Inst-BaseMerge-TIES | CombinHorizon/zetasepic-abliteratedV2-Qwen2.5-32B-Inst-BaseMerge-TIES | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8328136012446974}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764} |
HF Open LLM v2 | alibaba | CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES | dcd14b21-f2fd-4c10-bf83-b6bb946f2789 | 0.0.1 | hfopenllm_v2/CombinHorizon_Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES/1762652579.508495 | 1762652579.5084958 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES | CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8239958864701216}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.77} |
HF Open LLM v2 | alibaba | AtAndDev/Qwen2.5-1.5B-continuous-learnt | 1a2d8396-4ff1-4386-a76b-d4863c7736c5 | 0.0.1 | hfopenllm_v2/AtAndDev_Qwen2.5-1.5B-continuous-learnt/1762652579.483878 | 1762652579.4838789 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | AtAndDev/Qwen2.5-1.5B-continuous-learnt | AtAndDev/Qwen2.5-1.5B-continuous-learnt | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45105431366551857}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544} |
HF Open LLM v2 | alibaba | AtAndDev/Qwen2.5-1.5B-continuous-learnt | 4f7f368f-0646-4c16-80de-69d9c5e28193 | 0.0.1 | hfopenllm_v2/AtAndDev_Qwen2.5-1.5B-continuous-learnt/1762652579.483521 | 1762652579.483522 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | AtAndDev/Qwen2.5-1.5B-continuous-learnt | AtAndDev/Qwen2.5-1.5B-continuous-learnt | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4605214165081982}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544} |
HF Open LLM v2 | alibaba | DeepMount00/Qwen2-1.5B-Ita_v6 | 041f6e95-b7d1-44c6-a995-0c8257e188aa | 0.0.1 | hfopenllm_v2/DeepMount00_Qwen2-1.5B-Ita_v6/1762652579.553008 | 1762652579.5530088 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DeepMount00/Qwen2-1.5B-Ita_v6 | DeepMount00/Qwen2-1.5B-Ita_v6 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.29990425404593146}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.497} |
HF Open LLM v2 | alibaba | DeepMount00/Qwen2-1.5B-Ita_v5 | 04f0529b-474c-42d2-99a8-e3bdd5c18eaf | 0.0.1 | hfopenllm_v2/DeepMount00_Qwen2-1.5B-Ita_v5/1762652579.552789 | 1762652579.55279 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DeepMount00/Qwen2-1.5B-Ita_v5 | DeepMount00/Qwen2-1.5B-Ita_v5 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4987400098405564}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544} |
HF Open LLM v2 | alibaba | DeepMount00/Qwen2-1.5B-Ita_v3 | f9cac378-3bdb-4c66-8193-502773c5c5eb | 0.0.1 | hfopenllm_v2/DeepMount00_Qwen2-1.5B-Ita_v3/1762652579.552576 | 1762652579.552577 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DeepMount00/Qwen2-1.5B-Ita_v3 | DeepMount00/Qwen2-1.5B-Ita_v3 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4890479483326463}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544} |
HF Open LLM v2 | alibaba | DeepMount00/Qwen2-1.5B-Ita_v2 | 78ec8596-ee15-4e94-8bc8-77c6bdffc541 | 0.0.1 | hfopenllm_v2/DeepMount00_Qwen2-1.5B-Ita_v2/1762652579.552372 | 1762652579.552373 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DeepMount00/Qwen2-1.5B-Ita_v2 | DeepMount00/Qwen2-1.5B-Ita_v2 | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49998891829235315}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544} |
HF Open LLM v2 | alibaba | DeepMount00/Qwen2-1.5B-Ita | 6669c8b8-91d6-4f14-8cfb-a6422352850d | 0.0.1 | hfopenllm_v2/DeepMount00_Qwen2-1.5B-Ita/1762652579.5521228 | 1762652579.5521238 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DeepMount00/Qwen2-1.5B-Ita | DeepMount00/Qwen2-1.5B-Ita | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5173495214918638}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.544} |
HF Open LLM v2 | alibaba | hotmailuser/QwenSlerp-3B | 7f53fb66-2c19-434a-acec-7cdcf9fce04d | 0.0.1 | hfopenllm_v2/hotmailuser_QwenSlerp-3B/1762652580.1967301 | 1762652580.1967309 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenSlerp-3B | hotmailuser/QwenSlerp-3B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4333690164319561}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.397} |
HF Open LLM v2 | alibaba | hotmailuser/QwenSlerp3-14B | 7d2c1ffb-d1e7-4c88-af08-74642ddd8741 | 0.0.1 | hfopenllm_v2/hotmailuser_QwenSlerp3-14B/1762652580.197938 | 1762652580.1979399 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenSlerp3-14B | hotmailuser/QwenSlerp3-14B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6632291209546226}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | hotmailuser/QwenSlerp2-14B | 6732a278-0613-40fd-bdbc-88a586631279 | 0.0.1 | hfopenllm_v2/hotmailuser_QwenSlerp2-14B/1762652580.197355 | 1762652580.197356 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenSlerp2-14B | hotmailuser/QwenSlerp2-14B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7036707048409332}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | hotmailuser/QwenSlerp-7B | 4f8db3ee-409a-4bac-ab0a-ee3493d1e842 | 0.0.1 | hfopenllm_v2/hotmailuser_QwenSlerp-7B/1762652580.197109 | 1762652580.19711 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenSlerp-7B | hotmailuser/QwenSlerp-7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4672912317096415}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | hotmailuser/QwenStock1-14B | 67fd0572-cf55-412d-8ec6-0cb168d3ed08 | 0.0.1 | hfopenllm_v2/hotmailuser_QwenStock1-14B/1762652580.1990862 | 1762652580.1990871 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenStock1-14B | hotmailuser/QwenStock1-14B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6693240601603745}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | hotmailuser/QwenModelStock-1.8B | 661b1590-f312-447b-a494-1d37ffd93cae | 0.0.1 | hfopenllm_v2/hotmailuser_QwenModelStock-1.8B/1762652580.196316 | 1762652580.196316 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenModelStock-1.8B | hotmailuser/QwenModelStock-1.8B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3263075306852484}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777} |
HF Open LLM v2 | alibaba | hotmailuser/QwenStock-0.5B | 72853b4d-cc12-478f-b6f4-977b8fbabfa0 | 0.0.1 | hfopenllm_v2/hotmailuser_QwenStock-0.5B/1762652580.198598 | 1762652580.1985989 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenStock-0.5B | hotmailuser/QwenStock-0.5B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20490742341431845}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 0.63} |
HF Open LLM v2 | alibaba | hotmailuser/QwenStock-1.7B | 25674b98-92b5-4e2d-97ab-084eabb13db2 | 0.0.1 | hfopenllm_v2/hotmailuser_QwenStock-1.7B/1762652580.1988428 | 1762652580.198844 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenStock-1.7B | hotmailuser/QwenStock-1.7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.32141163224688274}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777} |
HF Open LLM v2 | alibaba | hotmailuser/QwenSlerp-14B | 83387977-a8cd-4cdd-abc7-301006380458 | 0.0.1 | hfopenllm_v2/hotmailuser_QwenSlerp-14B/1762652580.1965241 | 1762652580.196525 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenSlerp-14B | hotmailuser/QwenSlerp-14B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7024716640735471}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 14.766} |
HF Open LLM v2 | alibaba | hotmailuser/Deepseek-qwen-modelstock-2B | 15a4291f-4918-43a6-b242-90db88fe4a3d | 0.0.1 | hfopenllm_v2/hotmailuser_Deepseek-qwen-modelstock-2B/1762652580.1914759 | 1762652580.191477 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/Deepseek-qwen-modelstock-2B | hotmailuser/Deepseek-qwen-modelstock-2B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.21487431127186973}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.777} |
HF Open LLM v2 | alibaba | hotmailuser/QwenSlerp2-3B | cc53c4f9-3c1b-4b21-9aac-ea22dced76c3 | 0.0.1 | hfopenllm_v2/hotmailuser_QwenSlerp2-3B/1762652580.197566 | 1762652580.197566 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenSlerp2-3B | hotmailuser/QwenSlerp2-3B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4280486885907171}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.397} |
HF Open LLM v2 | alibaba | hotmailuser/Qwen2.5-HomerSlerp-7B | 9c7dab43-b26d-4cb4-a73c-95bb1e01ffe8 | 0.0.1 | hfopenllm_v2/hotmailuser_Qwen2.5-HomerSlerp-7B/1762652580.1961112 | 1762652580.1961112 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/Qwen2.5-HomerSlerp-7B | hotmailuser/Qwen2.5-HomerSlerp-7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44878145542715553}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | hotmailuser/QwenSparse-7B | 96bbc2c8-bb74-408d-8625-e6bf66b63cd0 | 0.0.1 | hfopenllm_v2/hotmailuser_QwenSparse-7B/1762652580.198252 | 1762652580.198254 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/QwenSparse-7B | hotmailuser/QwenSparse-7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.10858632871891026}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | adriszmar/QAIMath-Qwen2.5-7B-TIES | 78544e05-7eed-465d-9199-35b25e1bebfe | 0.0.1 | hfopenllm_v2/adriszmar_QAIMath-Qwen2.5-7B-TIES/1762652579.9747589 | 1762652579.9747598 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | adriszmar/QAIMath-Qwen2.5-7B-TIES | adriszmar/QAIMath-Qwen2.5-7B-TIES | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.174632198123202}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | adriszmar/QAIMath-Qwen2.5-7B-TIES | 457f0bc3-68e1-4ecb-a983-5f504b1246cd | 0.0.1 | hfopenllm_v2/adriszmar_QAIMath-Qwen2.5-7B-TIES/1762652579.975151 | 1762652579.975153 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | adriszmar/QAIMath-Qwen2.5-7B-TIES | adriszmar/QAIMath-Qwen2.5-7B-TIES | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.16853725891745014}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | alibaba | Qwen/QwQ-32B-Preview | 1326f0c0-9355-47ff-813b-0729370e1487 | 0.0.1 | hfopenllm_v2/Qwen_QwQ-32B-Preview/1762652579.834909 | 1762652579.83491 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/QwQ-32B-Preview | Qwen/QwQ-32B-Preview | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4035437084713006}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764} |
HF Open LLM v2 | alibaba | Qwen/QwQ-32B | 788241ad-d975-498e-80ef-b0d04bd8db85 | 0.0.1 | hfopenllm_v2/Qwen_QwQ-32B/1762652579.8346298 | 1762652579.834631 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/QwQ-32B | Qwen/QwQ-32B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.39767372793077926}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764} |
HF Open LLM v2 | alibaba | Qwen/Qwen1.5-7B | 102378fc-7b98-4088-a6f5-3039e7b638d5 | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-7B/1762652579.838115 | 1762652579.8381162 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-7B | Qwen/Qwen1.5-7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2684299879874289}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.721} |
HF Open LLM v2 | alibaba | Qwen/Qwen2.5-1.5B | 9982c576-75fd-47f6-8fe9-52b56fc58d3f | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-1.5B/1762652579.8426108 | 1762652579.842612 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-1.5B | Qwen/Qwen2.5-1.5B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.26743041795768563}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.5} |
HF Open LLM v2 | alibaba | Qwen/Qwen1.5-MoE-A2.7B | c6aa0ed8-3b79-4d73-8587-762e9469f4ce | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-MoE-A2.7B/1762652579.83854 | 1762652579.83854 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-MoE-A2.7B | Qwen/Qwen1.5-MoE-A2.7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.265982038768246}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2MoeForCausalLM", "params_billions": 14.316} |
HF Open LLM v2 | alibaba | Qwen/Qwen1.5-4B | 1e3f60f2-814a-4979-87bd-f5f94d5b09cc | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-4B/1762652579.837696 | 1762652579.837697 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-4B | Qwen/Qwen1.5-4B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24447466056729478}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.95} |
HF Open LLM v2 | alibaba | Qwen/Qwen1.5-32B | b8cd9221-dd4e-4f49-b03e-f11bdd5773e4 | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-32B/1762652579.837265 | 1762652579.837266 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-32B | Qwen/Qwen1.5-32B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.329729562006587}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.512} |
HF Open LLM v2 | alibaba | Qwen/Qwen1.5-1.8B | 7c828833-fd36-4a84-8530-d3c1769ca822 | 0.0.1 | hfopenllm_v2/Qwen_Qwen1.5-1.8B/1762652579.835954 | 1762652579.835955 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen1.5-1.8B | Qwen/Qwen1.5-1.8B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2154239639711521}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 1.837} |
HF Open LLM v2 | alibaba | Qwen/Qwen2.5-Math-7B | 8fddcebe-58d2-4d40-8147-f02feabc0d9c | 0.0.1 | hfopenllm_v2/Qwen_Qwen2.5-Math-7B/1762652579.8480499 | 1762652579.848052 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Qwen/Qwen2.5-Math-7B | Qwen/Qwen2.5-Math-7B | alibaba | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24599839536873275}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.