_leaderboard stringclasses 1
value | _developer stringclasses 559
values | _model stringlengths 9 102 | _uuid stringlengths 36 36 | schema_version stringclasses 1
value | evaluation_id stringlengths 35 133 | retrieved_timestamp stringlengths 13 18 | source_data stringclasses 1
value | evaluation_source_name stringclasses 1
value | evaluation_source_type stringclasses 1
value | source_organization_name stringclasses 1
value | source_organization_url null | source_organization_logo_url null | evaluator_relationship stringclasses 1
value | model_name stringlengths 4 102 | model_id stringlengths 9 102 | model_developer stringclasses 559
values | model_inference_platform stringclasses 1
value | evaluation_results stringlengths 1.35k 1.41k | additional_details stringclasses 660
values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
HF Open LLM v2 | google | HuggingFaceH4/zephyr-7b-gemma-v0.1 | dcf4d2bb-ee8f-4083-baf6-8870731515fa | 0.0.1 | hfopenllm_v2/HuggingFaceH4_zephyr-7b-gemma-v0.1/1762652579.641236 | 1762652579.641237 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | HuggingFaceH4/zephyr-7b-gemma-v0.1 | HuggingFaceH4/zephyr-7b-gemma-v0.1 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3363741539116212}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 8.538} |
HF Open LLM v2 | google | ehristoforu/Gemma2-9b-it-train6 | e289e629-17dd-440e-8839-d5dcbe535fd6 | 0.0.1 | hfopenllm_v2/ehristoforu_Gemma2-9b-it-train6/1762652580.1393359 | 1762652580.139337 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ehristoforu/Gemma2-9b-it-train6 | ehristoforu/Gemma2-9b-it-train6 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7025215317579578}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | ehristoforu/Gemma2-9B-it-psy10k-mental_health | 25c93024-ce65-49d5-96da-00107bb37f77 | 0.0.1 | hfopenllm_v2/ehristoforu_Gemma2-9B-it-psy10k-mental_health/1762652580.139083 | 1762652580.139084 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ehristoforu/Gemma2-9B-it-psy10k-mental_health | ehristoforu/Gemma2-9B-it-psy10k-mental_health | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5886658510529839}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | monsterapi/gemma-2-2b-LoRA-MonsterInstruct | f5395aa2-334b-410c-a2ee-4d7381f1c9bc | 0.0.1 | hfopenllm_v2/monsterapi_gemma-2-2b-LoRA-MonsterInstruct/1762652580.372597 | 1762652580.372598 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | monsterapi/gemma-2-2b-LoRA-MonsterInstruct | monsterapi/gemma-2-2b-LoRA-MonsterInstruct | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3902545246612322}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | cat-searcher/gemma-2-9b-it-sppo-iter-1-evol-1 | af7a7129-1b6a-4ff5-952f-075ae4f7c137 | 0.0.1 | hfopenllm_v2/cat-searcher_gemma-2-9b-it-sppo-iter-1-evol-1/1762652580.099224 | 1762652580.099225 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | cat-searcher/gemma-2-9b-it-sppo-iter-1-evol-1 | cat-searcher/gemma-2-9b-it-sppo-iter-1-evol-1 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2941827683878775}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | cat-searcher/gemma-2-9b-it-sppo-iter-1 | 3c33f6b0-dc40-4a61-bbbe-063b9d8d30e3 | 0.0.1 | hfopenllm_v2/cat-searcher_gemma-2-9b-it-sppo-iter-1/1762652580.091131 | 1762652580.091137 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | cat-searcher/gemma-2-9b-it-sppo-iter-1 | cat-searcher/gemma-2-9b-it-sppo-iter-1 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30147674836101546}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | google/umt5-base | 659053b0-7694-41e7-916d-28406b3ed572 | 0.0.1 | hfopenllm_v2/google_umt5-base/1762652580.180466 | 1762652580.180467 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/umt5-base | google/umt5-base | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.174632198123202}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "float16", "architecture": "UMT5ForConditionalGeneration", "params_billions": -1.0} |
HF Open LLM v2 | google | ifable/gemma-2-Ifable-9B | 42b3b64b-0e15-4a49-b542-da27ab7e2143 | 0.0.1 | hfopenllm_v2/ifable_gemma-2-Ifable-9B/1762652580.225604 | 1762652580.225605 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ifable/gemma-2-Ifable-9B | ifable/gemma-2-Ifable-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2984292787581395}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | openbmb | openbmb/MiniCPM-S-1B-sft-llama-format | 53b78e02-9491-4f3b-a03b-7c015dde640a | 0.0.1 | hfopenllm_v2/openbmb_MiniCPM-S-1B-sft-llama-format/1762652580.430347 | 1762652580.430348 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | openbmb/MiniCPM-S-1B-sft-llama-format | openbmb/MiniCPM-S-1B-sft-llama-format | openbmb | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3328767669782843}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.0} |
HF Open LLM v2 | EnnoAi | EnnoAi/EnnoAi-7B-French-Instruct-202502 | 75939d35-c0ca-4256-b667-fe6042ca5979 | 0.0.1 | hfopenllm_v2/EnnoAi_EnnoAi-7B-French-Instruct-202502/1762652579.596549 | 1762652579.59655 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | EnnoAi/EnnoAi-7B-French-Instruct-202502 | EnnoAi/EnnoAi-7B-French-Instruct-202502 | EnnoAi | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5564424615575562}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456} |
HF Open LLM v2 | CombinHorizon | CombinHorizon/YiSM-blossom5.1-34B-SLERP | 91ec4ba1-6948-48e8-8db0-a335b982c560 | 0.0.1 | hfopenllm_v2/CombinHorizon_YiSM-blossom5.1-34B-SLERP/1762652579.508977 | 1762652579.508977 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | CombinHorizon/YiSM-blossom5.1-34B-SLERP | CombinHorizon/YiSM-blossom5.1-34B-SLERP | CombinHorizon | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5033112142448702}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 34.389} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.0-2b-instruct | ec853cc1-7c48-4334-9ff6-d9669750570b | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.0-2b-instruct/1762652580.206777 | 1762652580.206777 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.0-2b-instruct | ibm-granite/granite-3.0-2b-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.513977357854936}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 2.634} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.1-3b-a800m-base | 8930e3f9-e0b8-4fb7-91e2-ee34b17cf1eb | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.1-3b-a800m-base/1762652580.20895 | 1762652580.208951 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.1-3b-a800m-base | ibm-granite/granite-3.1-3b-a800m-base | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2996294276962903}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GraniteMoeForCausalLM", "params_billions": 3.299} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.1-1b-a400m-instruct | 8167695b-db96-4687-91b8-0af55e67a606 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.1-1b-a400m-instruct/1762652580.208256 | 1762652580.208257 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.1-1b-a400m-instruct | ibm-granite/granite-3.1-1b-a400m-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46863987553025976}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "GraniteMoeForCausalLM", "params_billions": 1.335} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.0-2b-base | 184f8ef6-7cb7-45f2-b983-70dc4503a968 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.0-2b-base/1762652580.206552 | 1762652580.206552 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.0-2b-base | ibm-granite/granite-3.0-2b-base | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3873821460391761}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 2.634} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.0-3b-a800m-instruct | 7c92caf5-df83-4c8e-ab85-f99c7ac43f63 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.0-3b-a800m-instruct/1762652580.2071838 | 1762652580.2071848 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.0-3b-a800m-instruct | ibm-granite/granite-3.0-3b-a800m-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4298217618142085}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 3.374} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.1-3b-a800m-instruct | 1e0c27fc-8111-4325-8e61-c24c2f8124f7 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.1-3b-a800m-instruct/1762652580.2092001 | 1762652580.2092009 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.1-3b-a800m-instruct | ibm-granite/granite-3.1-3b-a800m-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5516462984880118}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "GraniteMoeForCausalLM", "params_billions": 3.299} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.1-8b-instruct | 6d6b2e81-8b90-4703-aafb-40de92b3ede3 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.1-8b-instruct/1762652580.2098079 | 1762652580.2098088 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.1-8b-instruct | ibm-granite/granite-3.1-8b-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7207564816908026}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "GraniteForCausalLM", "params_billions": 8.171} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.2-8b-instruct | 982accb5-ea5c-45bc-8cdd-08edf5e543a1 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.2-8b-instruct/1762652580.210291 | 1762652580.2102919 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.2-8b-instruct | ibm-granite/granite-3.2-8b-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7274509412802475}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 8.171} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.1-8b-base | 10cbee10-0344-4da0-a26a-4298fd8f4d11 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.1-8b-base/1762652580.209538 | 1762652580.2095392 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.1-8b-base | ibm-granite/granite-3.1-8b-base | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4221033524381973}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "GraniteForCausalLM", "params_billions": 8.171} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.0-3b-a800m-base | f917bdff-4be5-440b-8e62-bb9f7b0dd0f5 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.0-3b-a800m-base/1762652580.20698 | 1762652580.20698 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.0-3b-a800m-base | ibm-granite/granite-3.0-3b-a800m-base | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2732261510569733}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 3.374} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.0-1b-a400m-base | 52e253ba-0291-4e78-b292-806cabe74697 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.0-1b-a400m-base/1762652580.205958 | 1762652580.20596 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.0-1b-a400m-base | ibm-granite/granite-3.0-1b-a400m-base | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24040324117785256}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 1.335} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.2-2b-instruct | 39fd9dc4-88e4-4b52-8527-c1ea692d8ca1 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.2-2b-instruct/1762652580.2100549 | 1762652580.2100558 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.2-2b-instruct | ibm-granite/granite-3.2-2b-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6151688630611223}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 2.534} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.0-1b-a400m-instruct | afc49838-c7fc-40ed-841f-74b0bc3dd36e | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.0-1b-a400m-instruct/1762652580.206321 | 1762652580.206322 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.0-1b-a400m-instruct | ibm-granite/granite-3.0-1b-a400m-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.33315159332792543}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 1.335} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-7b-instruct | 509f5b3a-6110-4757-a313-80181ecd3228 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-7b-instruct/1762652580.2108219 | 1762652580.2108219 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-7b-instruct | ibm-granite/granite-7b-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2972313461615181}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 6.738} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.0-8b-instruct | d4dc4d78-33a3-428c-9490-382dd0c19c08 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.0-8b-instruct/1762652580.207594 | 1762652580.207595 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.0-8b-instruct | ibm-granite/granite-3.0-8b-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5309633993359841}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 8.171} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.1-1b-a400m-base | 17192714-a653-428d-a7c7-06dd41db77fa | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.1-1b-a400m-base/1762652580.207968 | 1762652580.2079701 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.1-1b-a400m-base | ibm-granite/granite-3.1-1b-a400m-base | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2519437315212525}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GraniteMoeForCausalLM", "params_billions": 1.335} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.1-2b-base | 971e6eba-61ff-42e6-9740-1895080ff94f | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.1-2b-base/1762652580.208491 | 1762652580.208492 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.1-2b-base | ibm-granite/granite-3.1-2b-base | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.35216115462528313}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 2.534} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-7b-base | 2d21a773-8f72-4b7d-ba94-80867127c54a | 0.0.1 | hfopenllm_v2/ibm-granite_granite-7b-base/1762652580.2106082 | 1762652580.210609 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-7b-base | ibm-granite/granite-7b-base | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24142719096441884}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 6.738} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.1-2b-instruct | fcdf14a1-900f-4856-aac6-8ed47910f882 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.1-2b-instruct/1762652580.2087219 | 1762652580.2087228 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.1-2b-instruct | ibm-granite/granite-3.1-2b-instruct | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.628557782240012}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "float16", "architecture": "GraniteForCausalLM", "params_billions": 2.534} |
HF Open LLM v2 | ibm-granite | ibm-granite/granite-3.0-8b-base | b7b71327-323b-4b7c-92a1-426911bed479 | 0.0.1 | hfopenllm_v2/ibm-granite_granite-3.0-8b-base/1762652580.207386 | 1762652580.207386 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ibm-granite/granite-3.0-8b-base | ibm-granite/granite-3.0-8b-base | ibm-granite | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4583482936386566}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 8.171} |
HF Open LLM v2 | Arthur-LAGACHERIE | Arthur-LAGACHERIE/Precis-1B-Instruct | d93c70b5-cb3b-4647-aa47-15c2401f5ebf | 0.0.1 | hfopenllm_v2/Arthur-LAGACHERIE_Precis-1B-Instruct/1762652579.482005 | 1762652579.482006 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Arthur-LAGACHERIE/Precis-1B-Instruct | Arthur-LAGACHERIE/Precis-1B-Instruct | Arthur-LAGACHERIE | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3670738086056109}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 1.236} |
HF Open LLM v2 | lunahr | lunahr/thea-v2-3b-50r | 03d675d8-ee8d-47de-8bf3-ef386bd8a88f | 0.0.1 | hfopenllm_v2/lunahr_thea-v2-3b-50r/1762652580.328458 | 1762652580.328459 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lunahr/thea-v2-3b-50r | lunahr/thea-v2-3b-50r | lunahr | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.370396104558128}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | lunahr | lunahr/thea-3b-50r-u1 | 977449d7-d8f0-4e32-b56c-8950006a09a4 | 0.0.1 | hfopenllm_v2/lunahr_thea-3b-50r-u1/1762652580.328209 | 1762652580.328209 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lunahr/thea-3b-50r-u1 | lunahr/thea-3b-50r-u1 | lunahr | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6030288523340293}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213} |
HF Open LLM v2 | dicta-il | dicta-il/dictalm2.0 | 613c1922-270a-4e8b-ae9d-20fa25573258 | 0.0.1 | hfopenllm_v2/dicta-il_dictalm2.0/1762652580.125907 | 1762652580.125909 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | dicta-il/dictalm2.0 | dicta-il/dictalm2.0 | dicta-il | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24132745559559746}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.251} |
HF Open LLM v2 | dicta-il | dicta-il/dictalm2.0-instruct | 4fc01471-7a04-4f46-a973-42f5a3fd67be | 0.0.1 | hfopenllm_v2/dicta-il_dictalm2.0-instruct/1762652580.126274 | 1762652580.126276 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | dicta-il/dictalm2.0-instruct | dicta-il/dictalm2.0-instruct | dicta-il | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44121264910437635}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.251} |
HF Open LLM v2 | shuttleai | shuttleai/shuttle-3 | bc357a38-215b-4885-9e0e-6f2b6f0bf1cc | 0.0.1 | hfopenllm_v2/shuttleai_shuttle-3/1762652580.5160902 | 1762652580.5160909 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | shuttleai/shuttle-3 | shuttleai/shuttle-3 | shuttleai | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.815403130360776}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706} |
HF Open LLM v2 | facebook | facebook/opt-30b | 1883ddb6-e4cc-4935-81ba-af30af1537e9 | 0.0.1 | hfopenllm_v2/facebook_opt-30b/1762652580.14943 | 1762652580.149431 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | facebook/opt-30b | facebook/opt-30b | facebook | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2452991396162183}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "OPTForCausalLM", "params_billions": 30.0} |
HF Open LLM v2 | facebook | facebook/opt-1.3b | 8675526d-af0b-4bf2-b143-123249371076 | 0.0.1 | hfopenllm_v2/facebook_opt-1.3b/1762652580.14919 | 1762652580.14919 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | facebook/opt-1.3b | facebook/opt-1.3b | facebook | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.23832985367713222}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "OPTForCausalLM", "params_billions": 1.3} |
HF Open LLM v2 | TheDrunkenSnail | TheDrunkenSnail/Mother-of-Rhodia-12B | 2178eb24-2558-44db-aff1-7903c2e0f657 | 0.0.1 | hfopenllm_v2/TheDrunkenSnail_Mother-of-Rhodia-12B/1762652579.9161909 | 1762652579.9161909 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | TheDrunkenSnail/Mother-of-Rhodia-12B | TheDrunkenSnail/Mother-of-Rhodia-12B | TheDrunkenSnail | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6504895898438365}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | TheDrunkenSnail | TheDrunkenSnail/Son-of-Rhodia | 22c87268-7e49-42b4-9bbb-16a4b305c595 | 0.0.1 | hfopenllm_v2/TheDrunkenSnail_Son-of-Rhodia/1762652579.916397 | 1762652579.916397 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | TheDrunkenSnail/Son-of-Rhodia | TheDrunkenSnail/Son-of-Rhodia | TheDrunkenSnail | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7046447869430887}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | TheDrunkenSnail | TheDrunkenSnail/Daughter-of-Rhodia-12B | 0f1c48a7-2a20-40c8-88e8-bdfdc3cdad40 | 0.0.1 | hfopenllm_v2/TheDrunkenSnail_Daughter-of-Rhodia-12B/1762652579.91594 | 1762652579.9159412 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | TheDrunkenSnail/Daughter-of-Rhodia-12B | TheDrunkenSnail/Daughter-of-Rhodia-12B | TheDrunkenSnail | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6903815210308648}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | Felladrin | Felladrin/Minueza-32M-UltraChat | 44324409-5cb3-438a-9751-9ee868b35233 | 0.0.1 | hfopenllm_v2/Felladrin_Minueza-32M-UltraChat/1762652579.6187 | 1762652579.6187022 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Felladrin/Minueza-32M-UltraChat | Felladrin/Minueza-32M-UltraChat | Felladrin | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.13756277787381924}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 0.033} |
HF Open LLM v2 | Felladrin | Felladrin/Llama-160M-Chat-v1 | 0885ef86-d7ef-4261-8ccd-f0391c42ffe4 | 0.0.1 | hfopenllm_v2/Felladrin_Llama-160M-Chat-v1/1762652579.618279 | 1762652579.61828 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Felladrin/Llama-160M-Chat-v1 | Felladrin/Llama-160M-Chat-v1 | Felladrin | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15754642127333254}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 0.162} |
HF Open LLM v2 | kavonalds | kavonalds/BunderMaxx-1010 | 6b0275ea-f2eb-4a37-922c-d1f734c1a6d3 | 0.0.1 | hfopenllm_v2/kavonalds_BunderMaxx-1010/1762652580.305197 | 1762652580.3051982 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | kavonalds/BunderMaxx-1010 | kavonalds/BunderMaxx-1010 | kavonalds | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2980558252104416}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.236} |
HF Open LLM v2 | kavonalds | kavonalds/BunderMaxx-0710 | 63d646bf-14d2-4cc7-ab82-efd1645cc1ba | 0.0.1 | hfopenllm_v2/kavonalds_BunderMaxx-0710/1762652580.3044312 | 1762652580.3044322 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | kavonalds/BunderMaxx-0710 | kavonalds/BunderMaxx-0710 | kavonalds | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.32825569488955975}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 1.236} |
HF Open LLM v2 | kavonalds | kavonalds/BunderMaxx-0710 | 10be7d08-18a9-43a6-80ea-81d704600eab | 0.0.1 | hfopenllm_v2/kavonalds_BunderMaxx-0710/1762652580.304877 | 1762652580.3048792 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | kavonalds/BunderMaxx-0710 | kavonalds/BunderMaxx-0710 | kavonalds | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.27007894608527594}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.236} |
HF Open LLM v2 | kavonalds | kavonalds/Lancer-1-1b-Instruct | ae2afa83-4607-43ea-be11-86cc57f3b848 | 0.0.1 | hfopenllm_v2/kavonalds_Lancer-1-1b-Instruct/1762652580.305463 | 1762652580.305465 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | kavonalds/Lancer-1-1b-Instruct | kavonalds/Lancer-1-1b-Instruct | kavonalds | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5545940327220664}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 1.236} |
HF Open LLM v2 | olabs-ai | olabs-ai/reflection_model | 84b63639-3343-4568-9fa7-d353ccb5b465 | 0.0.1 | hfopenllm_v2/olabs-ai_reflection_model/1762652580.417324 | 1762652580.417325 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | olabs-ai/reflection_model | olabs-ai/reflection_model | olabs-ai | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15986914719610634}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "?", "params_billions": 9.3} |
HF Open LLM v2 | UCLA-AGI | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3 | d8d05a10-8889-40aa-b56f-365e0a12052c | 0.0.1 | hfopenllm_v2/UCLA-AGI_Llama-3-Instruct-8B-SPPO-Iter3/1762652579.937166 | 1762652579.9371672 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3 | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3 | UCLA-AGI | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6834122350917787}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | UCLA-AGI | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3 | 49e095af-ed90-4e64-b476-4fc62d6e6997 | 0.0.1 | hfopenllm_v2/UCLA-AGI_Llama-3-Instruct-8B-SPPO-Iter3/1762652579.937367 | 1762652579.9373682 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3 | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3 | UCLA-AGI | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.67029814226253}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH",... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | UCLA-AGI | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter2 | 0c5c315f-63c4-427e-a307-1422a197895c | 0.0.1 | hfopenllm_v2/UCLA-AGI_Llama-3-Instruct-8B-SPPO-Iter2/1762652579.93697 | 1762652579.936971 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter2 | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter2 | UCLA-AGI | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6988745417713889}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.0} |
HF Open LLM v2 | UCLA-AGI | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter1 | 07af3512-a045-435e-a965-8daa0836905d | 0.0.1 | hfopenllm_v2/UCLA-AGI_Llama-3-Instruct-8B-SPPO-Iter1/1762652579.9367309 | 1762652579.9367318 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter1 | UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter1 | UCLA-AGI | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7298988904994304}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | DeepMount00 | DeepMount00/Lexora-Medium-7B | cc8f594a-e2f7-49e3-8654-57f1b397797f | 0.0.1 | hfopenllm_v2/DeepMount00_Lexora-Medium-7B/1762652579.551008 | 1762652579.551009 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DeepMount00/Lexora-Medium-7B | DeepMount00/Lexora-Medium-7B | DeepMount00 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4103379034295669}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | DeepMount00 | DeepMount00/Lexora-Lite-3B_v2 | bf38278f-6375-41a6-9744-04fb4a32ed72 | 0.0.1 | hfopenllm_v2/DeepMount00_Lexora-Lite-3B_v2/1762652579.550789 | 1762652579.550789 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DeepMount00/Lexora-Lite-3B_v2 | DeepMount00/Lexora-Lite-3B_v2 | DeepMount00 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49431840848947456}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.086} |
HF Open LLM v2 | DeepMount00 | DeepMount00/Lexora-Lite-3B | 5eb28bbd-8428-4385-b078-13e8a868e9f0 | 0.0.1 | hfopenllm_v2/DeepMount00_Lexora-Lite-3B/1762652579.550504 | 1762652579.550505 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DeepMount00/Lexora-Lite-3B | DeepMount00/Lexora-Lite-3B | DeepMount00 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5775996577968678}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.086} |
HF Open LLM v2 | DeepMount00 | DeepMount00/Qwen2.5-7B-Instruct-MathCoder | ea1a36fb-66c0-4b1a-bdac-7ec2602a7c65 | 0.0.1 | hfopenllm_v2/DeepMount00_Qwen2.5-7B-Instruct-MathCoder/1762652579.55323 | 1762652579.553231 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DeepMount00/Qwen2.5-7B-Instruct-MathCoder | DeepMount00/Qwen2.5-7B-Instruct-MathCoder | DeepMount00 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15302508455342934}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | DeepMount00 | DeepMount00/mergekit-ties-okvgjfz | 34350829-d42d-4e67-b23f-171044428c1f | 0.0.1 | hfopenllm_v2/DeepMount00_mergekit-ties-okvgjfz/1762652579.5535848 | 1762652579.553586 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DeepMount00/mergekit-ties-okvgjfz | DeepMount00/mergekit-ties-okvgjfz | DeepMount00 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15302508455342934}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616} |
HF Open LLM v2 | PrimeIntellect | PrimeIntellect/INTELLECT-1-Instruct | ea823c15-3c92-4a67-a4fd-7826a9dd9e41 | 0.0.1 | hfopenllm_v2/PrimeIntellect_INTELLECT-1-Instruct/1762652579.817848 | 1762652579.8178492 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | PrimeIntellect/INTELLECT-1-Instruct | PrimeIntellect/INTELLECT-1-Instruct | PrimeIntellect | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.0}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH", "lower_is_be... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.211} |
HF Open LLM v2 | PrimeIntellect | PrimeIntellect/INTELLECT-1 | fee7966f-3e1b-43d9-b129-b0c23aac53b5 | 0.0.1 | hfopenllm_v2/PrimeIntellect_INTELLECT-1/1762652579.817406 | 1762652579.817406 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | PrimeIntellect/INTELLECT-1 | PrimeIntellect/INTELLECT-1 | PrimeIntellect | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1757315035217667}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.211} |
HF Open LLM v2 | PrimeIntellect | PrimeIntellect/INTELLECT-1 | bfffc240-22ab-4cc0-97c8-466ddf472ac4 | 0.0.1 | hfopenllm_v2/PrimeIntellect_INTELLECT-1/1762652579.8176599 | 1762652579.817661 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | PrimeIntellect/INTELLECT-1 | PrimeIntellect/INTELLECT-1 | PrimeIntellect | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1757315035217667}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 10.211} |
HF Open LLM v2 | hotmailuser | hotmailuser/FalconSlerp-3B | 2db7aa3c-4969-40c0-b8c6-1ff5c953ba23 | 0.0.1 | hfopenllm_v2/hotmailuser_FalconSlerp-3B/1762652580.19236 | 1762652580.1923609 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/FalconSlerp-3B | hotmailuser/FalconSlerp-3B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5694568190179834}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.228} |
HF Open LLM v2 | hotmailuser | hotmailuser/Falcon3Slerp1-10B | 376d342c-669b-4c76-9e7b-d49566ac441d | 0.0.1 | hfopenllm_v2/hotmailuser_Falcon3Slerp1-10B/1762652580.19171 | 1762652580.191711 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/Falcon3Slerp1-10B | hotmailuser/Falcon3Slerp1-10B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5694069513335727}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.306} |
HF Open LLM v2 | hotmailuser | hotmailuser/RombosBeagle-v2beta-MGS-32B | c507c0ac-759a-4013-8dd0-7ab5a959ca65 | 0.0.1 | hfopenllm_v2/hotmailuser_RombosBeagle-v2beta-MGS-32B/1762652580.199307 | 1762652580.199308 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/RombosBeagle-v2beta-MGS-32B | hotmailuser/RombosBeagle-v2beta-MGS-32B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5156761836371937}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764} |
HF Open LLM v2 | hotmailuser | hotmailuser/FalconSlerp3-7B | 017a681e-1bbb-4890-bfcc-f276954678e1 | 0.0.1 | hfopenllm_v2/hotmailuser_FalconSlerp3-7B/1762652580.193249 | 1762652580.19325 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/FalconSlerp3-7B | hotmailuser/FalconSlerp3-7B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6096235765546527}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456} |
HF Open LLM v2 | hotmailuser | hotmailuser/Falcon3Slerp2-10B | bae0b772-8ae6-4fed-ae78-d6d83e560a95 | 0.0.1 | hfopenllm_v2/hotmailuser_Falcon3Slerp2-10B/1762652580.191951 | 1762652580.191952 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/Falcon3Slerp2-10B | hotmailuser/Falcon3Slerp2-10B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6117966994241945}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.306} |
HF Open LLM v2 | hotmailuser | hotmailuser/Falcon3Slerp4-10B | d5466af4-2bef-4ce8-a659-9e05a5e674b6 | 0.0.1 | hfopenllm_v2/hotmailuser_Falcon3Slerp4-10B/1762652580.19215 | 1762652580.192151 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/Falcon3Slerp4-10B | hotmailuser/Falcon3Slerp4-10B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6072254950198805}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.306} |
HF Open LLM v2 | hotmailuser | hotmailuser/FalconSlerp2-7B | fc8605ad-f7b9-4a73-afd3-85b996fc2549 | 0.0.1 | hfopenllm_v2/hotmailuser_FalconSlerp2-7B/1762652580.1928341 | 1762652580.192835 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/FalconSlerp2-7B | hotmailuser/FalconSlerp2-7B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6160432097944565}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456} |
HF Open LLM v2 | hotmailuser | hotmailuser/FalconSlerp3-10B | f933fbc2-370e-4231-94a9-c833c2aa793d | 0.0.1 | hfopenllm_v2/hotmailuser_FalconSlerp3-10B/1762652580.1930392 | 1762652580.19304 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/FalconSlerp3-10B | hotmailuser/FalconSlerp3-10B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6001564737119731}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 10.306} |
HF Open LLM v2 | hotmailuser | hotmailuser/FalconSlerp1-7B | 5d01fa6d-4280-4926-b166-e98892ee60f4 | 0.0.1 | hfopenllm_v2/hotmailuser_FalconSlerp1-7B/1762652580.1925812 | 1762652580.192582 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/FalconSlerp1-7B | hotmailuser/FalconSlerp1-7B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5394564200765082}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456} |
HF Open LLM v2 | hotmailuser | hotmailuser/FalconSlerp6-7B | 88a4587f-d3d4-4b08-b800-13a2daf4a660 | 0.0.1 | hfopenllm_v2/hotmailuser_FalconSlerp6-7B/1762652580.193665 | 1762652580.193666 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/FalconSlerp6-7B | hotmailuser/FalconSlerp6-7B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6026542906155667}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456} |
HF Open LLM v2 | hotmailuser | hotmailuser/FalconSlerp4-7B | d6ac7c9f-212e-4000-b89e-d977122d2e2b | 0.0.1 | hfopenllm_v2/hotmailuser_FalconSlerp4-7B/1762652580.193457 | 1762652580.1934578 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/FalconSlerp4-7B | hotmailuser/FalconSlerp4-7B | hotmailuser | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6284580468711907}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456} |
HF Open LLM v2 | Solshine | Solshine/Brimful-merged-replete | 6523a08c-7a43-4784-9650-e1d5144fcfcf | 0.0.1 | hfopenllm_v2/Solshine_Brimful-merged-replete/1762652579.8890932 | 1762652579.8890939 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Solshine/Brimful-merged-replete | Solshine/Brimful-merged-replete | Solshine | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.17605619755581856}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 12.277} |
HF Open LLM v2 | oobabooga | oobabooga/CodeBooga-34B-v0.1 | 3086045f-e22d-4aca-9459-fc64454a2fb2 | 0.0.1 | hfopenllm_v2/oobabooga_CodeBooga-34B-v0.1/1762652580.425838 | 1762652580.425838 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | oobabooga/CodeBooga-34B-v0.1 | oobabooga/CodeBooga-34B-v0.1 | oobabooga | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5250180631834643}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 33.744} |
HF Open LLM v2 | gaverfraxz | gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES | 83a638be-6f3d-4d5b-b1de-6515634aebbd | 0.0.1 | hfopenllm_v2/gaverfraxz_Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES/1762652580.163549 | 1762652580.16355 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES | gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES | gaverfraxz | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45505148561372716}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | gaverfraxz | gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-DELLA | 3666aa17-279d-4f0b-a6c2-2c8198729df9 | 0.0.1 | hfopenllm_v2/gaverfraxz_Meta-Llama-3.1-8B-Instruct-HalfAbliterated-DELLA/1762652580.163272 | 1762652580.1632729 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-DELLA | gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-DELLA | gaverfraxz | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.40094615619888563}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | crestf411 | crestf411/MN-Slush | b32a7808-7a64-41a8-aad4-030efc512906 | 0.0.1 | hfopenllm_v2/crestf411_MN-Slush/1762652580.117737 | 1762652580.117738 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | crestf411/MN-Slush | crestf411/MN-Slush | crestf411 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4077148632295642}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | Invalid-Null | Invalid-Null/PeiYangMe-0.7 | 294c1745-38cb-4b1e-aae6-e2878ab9065a | 0.0.1 | hfopenllm_v2/Invalid-Null_PeiYangMe-0.7/1762652579.648521 | 1762652579.648522 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Invalid-Null/PeiYangMe-0.7 | Invalid-Null/PeiYangMe-0.7 | Invalid-Null | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1491032682172192}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 6.061} |
HF Open LLM v2 | Invalid-Null | Invalid-Null/PeiYangMe-0.5 | c645a252-366a-4890-a16b-bf687bfbb593 | 0.0.1 | hfopenllm_v2/Invalid-Null_PeiYangMe-0.5/1762652579.648252 | 1762652579.648252 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Invalid-Null/PeiYangMe-0.5 | Invalid-Null/PeiYangMe-0.5 | Invalid-Null | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14088507382633633}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 6.061} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-28 | 0dc95982-e5b0-4011-9e5b-48af7e3048f0 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-28/1762652580.252297 | 1762652580.2522979 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-28 | jaspionjader/bh-28 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4785070280189896}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-11 | 804f4be8-a8a9-473f-a898-d71b742a62eb | 0.0.1 | hfopenllm_v2/jaspionjader_bh-11/1762652580.2481 | 1762652580.2481012 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-11 | jaspionjader/bh-11 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45754930251732073}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-13 | 98772920-a700-4fda-88fd-53c16ac4b1a1 | 0.0.1 | hfopenllm_v2/jaspionjader_test-13/1762652580.278408 | 1762652580.278409 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-13 | jaspionjader/test-13 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45809895521660304}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v5-8B | 53c89eb1-49ab-4e5f-b1ad-d8e80045a292 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v5-8B/1762652580.2420359 | 1762652580.2420359 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v5-8B | jaspionjader/Kosmos-EVAA-v5-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44595894448951}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH",... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/kstc-5-8b | ea79ca75-c55b-457a-b952-528a22567dbb | 0.0.1 | hfopenllm_v2/jaspionjader_kstc-5-8b/1762652580.270952 | 1762652580.270953 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/kstc-5-8b | jaspionjader/kstc-5-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47208739477918593}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v4-8B | fd2a2a9c-639f-4348-9861-00271ed070b2 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v4-8B/1762652580.241815 | 1762652580.241816 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v4-8B | jaspionjader/Kosmos-EVAA-v4-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4289230353240513}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-8 | f6dced28-f64c-4995-88b1-ac9a82903de2 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-8/1762652580.262149 | 1762652580.262152 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-8 | jaspionjader/bh-8 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45967255770245175}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/f-4-8b | a98ec95c-4af0-4b55-adbc-06e5ceecd00f | 0.0.1 | hfopenllm_v2/jaspionjader_f-4-8b/1762652580.265391 | 1762652580.2653928 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/f-4-8b | jaspionjader/f-4-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4797060687863757}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Auro-Kosmos-EVAA-v2.2-8B | 4e616fc6-8baa-4c9a-9098-b8d108911ad2 | 0.0.1 | hfopenllm_v2/jaspionjader_Auro-Kosmos-EVAA-v2.2-8B/1762652580.231466 | 1762652580.231467 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Auro-Kosmos-EVAA-v2.2-8B | jaspionjader/Auro-Kosmos-EVAA-v2.2-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4267997801389203}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-16 | 80c756a7-9d47-4b49-bf42-bbada0909163 | 0.0.1 | hfopenllm_v2/jaspionjader_test-16/1762652580.279189 | 1762652580.27919 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-16 | jaspionjader/test-16 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4599473840520929}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
HF Open LLM v2 | jaspionjader | jaspionjader/fr-10-8b | 725e5a72-548f-46d0-b268-12209e5cb085 | 0.0.1 | hfopenllm_v2/jaspionjader_fr-10-8b/1762652580.268136 | 1762652580.268136 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/fr-10-8b | jaspionjader/fr-10-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44018869931781013}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/f-3-8b | 5ba1e4d3-29d4-4337-bd10-9e1a5df29af4 | 0.0.1 | hfopenllm_v2/jaspionjader_f-3-8b/1762652580.264997 | 1762652580.264998 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/f-3-8b | jaspionjader/f-3-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4803055891700687}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-8B | 746ffa2c-cc95-4d69-9e46-0e8f4febd440 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-8B/1762652580.232065 | 1762652580.232065 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-8B | jaspionjader/Kosmos-EVAA-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4404635256674513}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-55 | 7c388cc5-fb2f-48ba-967c-a931fcb25a42 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-55/1762652580.259115 | 1762652580.259116 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-55 | jaspionjader/bh-55 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47093822169621047}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-2 | 3e4b8dcc-9270-4b14-952f-c6b96ee8ce57 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-2/1762652580.250077 | 1762652580.250078 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-2 | jaspionjader/bh-2 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45792386423578324}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-v14-8B | c20f5702-24fc-443a-875e-495401776eeb | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-v14-8B/1762652580.23884 | 1762652580.23884 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-v14-8B | jaspionjader/Kosmos-EVAA-gamma-v14-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4380155764482684}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-14 | d647b482-3d3b-4ed4-b8b5-d57eedf87db9 | 0.0.1 | hfopenllm_v2/jaspionjader_test-14/1762652580.2787268 | 1762652580.278728 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-14 | jaspionjader/test-14 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4443853420036614}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
HF Open LLM v2 | jaspionjader | jaspionjader/PRP-Kosmos-EVAA-8B | 84a37d06-2668-4143-8e2f-5a08651f2dfb | 0.0.1 | hfopenllm_v2/jaspionjader_PRP-Kosmos-EVAA-8B/1762652580.244709 | 1762652580.24471 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/PRP-Kosmos-EVAA-8B | jaspionjader/PRP-Kosmos-EVAA-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.36327721556580983}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-10 | 86411dbb-e28b-4e9d-856e-fcc001252fbe | 0.0.1 | hfopenllm_v2/jaspionjader_bh-10/1762652580.247846 | 1762652580.2478468 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-10 | jaspionjader/bh-10 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46184568057199343}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-4 | 85ba493b-05f1-4853-a0ff-44570a7c2a82 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-4/1762652580.2554429 | 1762652580.255444 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-4 | jaspionjader/bh-4 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4672912317096415}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-ultra-light-8B | 1810feae-7a27-4c17-8174-3cd8a143b21f | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-ultra-light-8B/1762652580.238316 | 1762652580.238317 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-ultra-light-8B | jaspionjader/Kosmos-EVAA-gamma-ultra-light-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4563003940655239}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.