_leaderboard
stringclasses
1 value
_developer
stringclasses
559 values
_model
stringlengths
9
102
_uuid
stringlengths
36
36
schema_version
stringclasses
1 value
evaluation_id
stringlengths
35
133
retrieved_timestamp
stringlengths
13
18
source_data
stringclasses
1 value
evaluation_source_name
stringclasses
1 value
evaluation_source_type
stringclasses
1 value
source_organization_name
stringclasses
1 value
source_organization_url
null
source_organization_logo_url
null
evaluator_relationship
stringclasses
1 value
model_name
stringlengths
4
102
model_id
stringlengths
9
102
model_developer
stringclasses
559 values
model_inference_platform
stringclasses
1 value
evaluation_results
stringlengths
1.35k
1.41k
additional_details
stringclasses
660 values
HF Open LLM v2
google
HuggingFaceH4/zephyr-7b-gemma-v0.1
dcf4d2bb-ee8f-4083-baf6-8870731515fa
0.0.1
hfopenllm_v2/HuggingFaceH4_zephyr-7b-gemma-v0.1/1762652579.641236
1762652579.641237
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
HuggingFaceH4/zephyr-7b-gemma-v0.1
HuggingFaceH4/zephyr-7b-gemma-v0.1
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3363741539116212}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 8.538}
HF Open LLM v2
google
ehristoforu/Gemma2-9b-it-train6
e289e629-17dd-440e-8839-d5dcbe535fd6
0.0.1
hfopenllm_v2/ehristoforu_Gemma2-9b-it-train6/1762652580.1393359
1762652580.139337
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ehristoforu/Gemma2-9b-it-train6
ehristoforu/Gemma2-9b-it-train6
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7025215317579578}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
ehristoforu/Gemma2-9B-it-psy10k-mental_health
25c93024-ce65-49d5-96da-00107bb37f77
0.0.1
hfopenllm_v2/ehristoforu_Gemma2-9B-it-psy10k-mental_health/1762652580.139083
1762652580.139084
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ehristoforu/Gemma2-9B-it-psy10k-mental_health
ehristoforu/Gemma2-9B-it-psy10k-mental_health
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5886658510529839}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
monsterapi/gemma-2-2b-LoRA-MonsterInstruct
f5395aa2-334b-410c-a2ee-4d7381f1c9bc
0.0.1
hfopenllm_v2/monsterapi_gemma-2-2b-LoRA-MonsterInstruct/1762652580.372597
1762652580.372598
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
monsterapi/gemma-2-2b-LoRA-MonsterInstruct
monsterapi/gemma-2-2b-LoRA-MonsterInstruct
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3902545246612322}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
cat-searcher/gemma-2-9b-it-sppo-iter-1-evol-1
af7a7129-1b6a-4ff5-952f-075ae4f7c137
0.0.1
hfopenllm_v2/cat-searcher_gemma-2-9b-it-sppo-iter-1-evol-1/1762652580.099224
1762652580.099225
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
cat-searcher/gemma-2-9b-it-sppo-iter-1-evol-1
cat-searcher/gemma-2-9b-it-sppo-iter-1-evol-1
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2941827683878775}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
cat-searcher/gemma-2-9b-it-sppo-iter-1
3c33f6b0-dc40-4a61-bbbe-063b9d8d30e3
0.0.1
hfopenllm_v2/cat-searcher_gemma-2-9b-it-sppo-iter-1/1762652580.091131
1762652580.091137
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
cat-searcher/gemma-2-9b-it-sppo-iter-1
cat-searcher/gemma-2-9b-it-sppo-iter-1
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30147674836101546}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
google/umt5-base
659053b0-7694-41e7-916d-28406b3ed572
0.0.1
hfopenllm_v2/google_umt5-base/1762652580.180466
1762652580.180467
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/umt5-base
google/umt5-base
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.174632198123202}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "float16", "architecture": "UMT5ForConditionalGeneration", "params_billions": -1.0}
HF Open LLM v2
google
ifable/gemma-2-Ifable-9B
42b3b64b-0e15-4a49-b542-da27ab7e2143
0.0.1
hfopenllm_v2/ifable_gemma-2-Ifable-9B/1762652580.225604
1762652580.225605
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ifable/gemma-2-Ifable-9B
ifable/gemma-2-Ifable-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2984292787581395}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
openbmb
openbmb/MiniCPM-S-1B-sft-llama-format
53b78e02-9491-4f3b-a03b-7c015dde640a
0.0.1
hfopenllm_v2/openbmb_MiniCPM-S-1B-sft-llama-format/1762652580.430347
1762652580.430348
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
openbmb/MiniCPM-S-1B-sft-llama-format
openbmb/MiniCPM-S-1B-sft-llama-format
openbmb
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3328767669782843}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.0}
HF Open LLM v2
EnnoAi
EnnoAi/EnnoAi-7B-French-Instruct-202502
75939d35-c0ca-4256-b667-fe6042ca5979
0.0.1
hfopenllm_v2/EnnoAi_EnnoAi-7B-French-Instruct-202502/1762652579.596549
1762652579.59655
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
EnnoAi/EnnoAi-7B-French-Instruct-202502
EnnoAi/EnnoAi-7B-French-Instruct-202502
EnnoAi
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5564424615575562}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456}
HF Open LLM v2
CombinHorizon
CombinHorizon/YiSM-blossom5.1-34B-SLERP
91ec4ba1-6948-48e8-8db0-a335b982c560
0.0.1
hfopenllm_v2/CombinHorizon_YiSM-blossom5.1-34B-SLERP/1762652579.508977
1762652579.508977
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
CombinHorizon/YiSM-blossom5.1-34B-SLERP
CombinHorizon/YiSM-blossom5.1-34B-SLERP
CombinHorizon
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5033112142448702}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 34.389}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.0-2b-instruct
ec853cc1-7c48-4334-9ff6-d9669750570b
0.0.1
hfopenllm_v2/ibm-granite_granite-3.0-2b-instruct/1762652580.206777
1762652580.206777
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.0-2b-instruct
ibm-granite/granite-3.0-2b-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.513977357854936}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 2.634}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.1-3b-a800m-base
8930e3f9-e0b8-4fb7-91e2-ee34b17cf1eb
0.0.1
hfopenllm_v2/ibm-granite_granite-3.1-3b-a800m-base/1762652580.20895
1762652580.208951
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.1-3b-a800m-base
ibm-granite/granite-3.1-3b-a800m-base
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2996294276962903}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GraniteMoeForCausalLM", "params_billions": 3.299}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.1-1b-a400m-instruct
8167695b-db96-4687-91b8-0af55e67a606
0.0.1
hfopenllm_v2/ibm-granite_granite-3.1-1b-a400m-instruct/1762652580.208256
1762652580.208257
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.1-1b-a400m-instruct
ibm-granite/granite-3.1-1b-a400m-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46863987553025976}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "GraniteMoeForCausalLM", "params_billions": 1.335}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.0-2b-base
184f8ef6-7cb7-45f2-b983-70dc4503a968
0.0.1
hfopenllm_v2/ibm-granite_granite-3.0-2b-base/1762652580.206552
1762652580.206552
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.0-2b-base
ibm-granite/granite-3.0-2b-base
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3873821460391761}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 2.634}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.0-3b-a800m-instruct
7c92caf5-df83-4c8e-ab85-f99c7ac43f63
0.0.1
hfopenllm_v2/ibm-granite_granite-3.0-3b-a800m-instruct/1762652580.2071838
1762652580.2071848
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.0-3b-a800m-instruct
ibm-granite/granite-3.0-3b-a800m-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4298217618142085}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 3.374}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.1-3b-a800m-instruct
1e0c27fc-8111-4325-8e61-c24c2f8124f7
0.0.1
hfopenllm_v2/ibm-granite_granite-3.1-3b-a800m-instruct/1762652580.2092001
1762652580.2092009
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.1-3b-a800m-instruct
ibm-granite/granite-3.1-3b-a800m-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5516462984880118}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "GraniteMoeForCausalLM", "params_billions": 3.299}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.1-8b-instruct
6d6b2e81-8b90-4703-aafb-40de92b3ede3
0.0.1
hfopenllm_v2/ibm-granite_granite-3.1-8b-instruct/1762652580.2098079
1762652580.2098088
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.1-8b-instruct
ibm-granite/granite-3.1-8b-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7207564816908026}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "GraniteForCausalLM", "params_billions": 8.171}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.2-8b-instruct
982accb5-ea5c-45bc-8cdd-08edf5e543a1
0.0.1
hfopenllm_v2/ibm-granite_granite-3.2-8b-instruct/1762652580.210291
1762652580.2102919
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.2-8b-instruct
ibm-granite/granite-3.2-8b-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7274509412802475}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 8.171}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.1-8b-base
10cbee10-0344-4da0-a26a-4298fd8f4d11
0.0.1
hfopenllm_v2/ibm-granite_granite-3.1-8b-base/1762652580.209538
1762652580.2095392
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.1-8b-base
ibm-granite/granite-3.1-8b-base
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4221033524381973}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "GraniteForCausalLM", "params_billions": 8.171}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.0-3b-a800m-base
f917bdff-4be5-440b-8e62-bb9f7b0dd0f5
0.0.1
hfopenllm_v2/ibm-granite_granite-3.0-3b-a800m-base/1762652580.20698
1762652580.20698
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.0-3b-a800m-base
ibm-granite/granite-3.0-3b-a800m-base
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2732261510569733}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 3.374}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.0-1b-a400m-base
52e253ba-0291-4e78-b292-806cabe74697
0.0.1
hfopenllm_v2/ibm-granite_granite-3.0-1b-a400m-base/1762652580.205958
1762652580.20596
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.0-1b-a400m-base
ibm-granite/granite-3.0-1b-a400m-base
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24040324117785256}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 1.335}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.2-2b-instruct
39fd9dc4-88e4-4b52-8527-c1ea692d8ca1
0.0.1
hfopenllm_v2/ibm-granite_granite-3.2-2b-instruct/1762652580.2100549
1762652580.2100558
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.2-2b-instruct
ibm-granite/granite-3.2-2b-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6151688630611223}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 2.534}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.0-1b-a400m-instruct
afc49838-c7fc-40ed-841f-74b0bc3dd36e
0.0.1
hfopenllm_v2/ibm-granite_granite-3.0-1b-a400m-instruct/1762652580.206321
1762652580.206322
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.0-1b-a400m-instruct
ibm-granite/granite-3.0-1b-a400m-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.33315159332792543}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 1.335}
HF Open LLM v2
ibm-granite
ibm-granite/granite-7b-instruct
509f5b3a-6110-4757-a313-80181ecd3228
0.0.1
hfopenllm_v2/ibm-granite_granite-7b-instruct/1762652580.2108219
1762652580.2108219
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-7b-instruct
ibm-granite/granite-7b-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2972313461615181}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 6.738}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.0-8b-instruct
d4dc4d78-33a3-428c-9490-382dd0c19c08
0.0.1
hfopenllm_v2/ibm-granite_granite-3.0-8b-instruct/1762652580.207594
1762652580.207595
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.0-8b-instruct
ibm-granite/granite-3.0-8b-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5309633993359841}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 8.171}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.1-1b-a400m-base
17192714-a653-428d-a7c7-06dd41db77fa
0.0.1
hfopenllm_v2/ibm-granite_granite-3.1-1b-a400m-base/1762652580.207968
1762652580.2079701
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.1-1b-a400m-base
ibm-granite/granite-3.1-1b-a400m-base
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2519437315212525}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GraniteMoeForCausalLM", "params_billions": 1.335}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.1-2b-base
971e6eba-61ff-42e6-9740-1895080ff94f
0.0.1
hfopenllm_v2/ibm-granite_granite-3.1-2b-base/1762652580.208491
1762652580.208492
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.1-2b-base
ibm-granite/granite-3.1-2b-base
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.35216115462528313}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 2.534}
HF Open LLM v2
ibm-granite
ibm-granite/granite-7b-base
2d21a773-8f72-4b7d-ba94-80867127c54a
0.0.1
hfopenllm_v2/ibm-granite_granite-7b-base/1762652580.2106082
1762652580.210609
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-7b-base
ibm-granite/granite-7b-base
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24142719096441884}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 6.738}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.1-2b-instruct
fcdf14a1-900f-4856-aac6-8ed47910f882
0.0.1
hfopenllm_v2/ibm-granite_granite-3.1-2b-instruct/1762652580.2087219
1762652580.2087228
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.1-2b-instruct
ibm-granite/granite-3.1-2b-instruct
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.628557782240012}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "float16", "architecture": "GraniteForCausalLM", "params_billions": 2.534}
HF Open LLM v2
ibm-granite
ibm-granite/granite-3.0-8b-base
b7b71327-323b-4b7c-92a1-426911bed479
0.0.1
hfopenllm_v2/ibm-granite_granite-3.0-8b-base/1762652580.207386
1762652580.207386
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ibm-granite/granite-3.0-8b-base
ibm-granite/granite-3.0-8b-base
ibm-granite
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4583482936386566}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GraniteForCausalLM", "params_billions": 8.171}
HF Open LLM v2
Arthur-LAGACHERIE
Arthur-LAGACHERIE/Precis-1B-Instruct
d93c70b5-cb3b-4647-aa47-15c2401f5ebf
0.0.1
hfopenllm_v2/Arthur-LAGACHERIE_Precis-1B-Instruct/1762652579.482005
1762652579.482006
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Arthur-LAGACHERIE/Precis-1B-Instruct
Arthur-LAGACHERIE/Precis-1B-Instruct
Arthur-LAGACHERIE
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3670738086056109}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 1.236}
HF Open LLM v2
lunahr
lunahr/thea-v2-3b-50r
03d675d8-ee8d-47de-8bf3-ef386bd8a88f
0.0.1
hfopenllm_v2/lunahr_thea-v2-3b-50r/1762652580.328458
1762652580.328459
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lunahr/thea-v2-3b-50r
lunahr/thea-v2-3b-50r
lunahr
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.370396104558128}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
lunahr
lunahr/thea-3b-50r-u1
977449d7-d8f0-4e32-b56c-8950006a09a4
0.0.1
hfopenllm_v2/lunahr_thea-3b-50r-u1/1762652580.328209
1762652580.328209
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lunahr/thea-3b-50r-u1
lunahr/thea-3b-50r-u1
lunahr
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6030288523340293}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.213}
HF Open LLM v2
dicta-il
dicta-il/dictalm2.0
613c1922-270a-4e8b-ae9d-20fa25573258
0.0.1
hfopenllm_v2/dicta-il_dictalm2.0/1762652580.125907
1762652580.125909
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
dicta-il/dictalm2.0
dicta-il/dictalm2.0
dicta-il
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24132745559559746}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.251}
HF Open LLM v2
dicta-il
dicta-il/dictalm2.0-instruct
4fc01471-7a04-4f46-a973-42f5a3fd67be
0.0.1
hfopenllm_v2/dicta-il_dictalm2.0-instruct/1762652580.126274
1762652580.126276
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
dicta-il/dictalm2.0-instruct
dicta-il/dictalm2.0-instruct
dicta-il
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44121264910437635}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.251}
HF Open LLM v2
shuttleai
shuttleai/shuttle-3
bc357a38-215b-4885-9e0e-6f2b6f0bf1cc
0.0.1
hfopenllm_v2/shuttleai_shuttle-3/1762652580.5160902
1762652580.5160909
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
shuttleai/shuttle-3
shuttleai/shuttle-3
shuttleai
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.815403130360776}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 72.706}
HF Open LLM v2
facebook
facebook/opt-30b
1883ddb6-e4cc-4935-81ba-af30af1537e9
0.0.1
hfopenllm_v2/facebook_opt-30b/1762652580.14943
1762652580.149431
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
facebook/opt-30b
facebook/opt-30b
facebook
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2452991396162183}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "OPTForCausalLM", "params_billions": 30.0}
HF Open LLM v2
facebook
facebook/opt-1.3b
8675526d-af0b-4bf2-b143-123249371076
0.0.1
hfopenllm_v2/facebook_opt-1.3b/1762652580.14919
1762652580.14919
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
facebook/opt-1.3b
facebook/opt-1.3b
facebook
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.23832985367713222}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "OPTForCausalLM", "params_billions": 1.3}
HF Open LLM v2
TheDrunkenSnail
TheDrunkenSnail/Mother-of-Rhodia-12B
2178eb24-2558-44db-aff1-7903c2e0f657
0.0.1
hfopenllm_v2/TheDrunkenSnail_Mother-of-Rhodia-12B/1762652579.9161909
1762652579.9161909
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
TheDrunkenSnail/Mother-of-Rhodia-12B
TheDrunkenSnail/Mother-of-Rhodia-12B
TheDrunkenSnail
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6504895898438365}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
TheDrunkenSnail
TheDrunkenSnail/Son-of-Rhodia
22c87268-7e49-42b4-9bbb-16a4b305c595
0.0.1
hfopenllm_v2/TheDrunkenSnail_Son-of-Rhodia/1762652579.916397
1762652579.916397
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
TheDrunkenSnail/Son-of-Rhodia
TheDrunkenSnail/Son-of-Rhodia
TheDrunkenSnail
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7046447869430887}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
TheDrunkenSnail
TheDrunkenSnail/Daughter-of-Rhodia-12B
0f1c48a7-2a20-40c8-88e8-bdfdc3cdad40
0.0.1
hfopenllm_v2/TheDrunkenSnail_Daughter-of-Rhodia-12B/1762652579.91594
1762652579.9159412
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
TheDrunkenSnail/Daughter-of-Rhodia-12B
TheDrunkenSnail/Daughter-of-Rhodia-12B
TheDrunkenSnail
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6903815210308648}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
Felladrin
Felladrin/Minueza-32M-UltraChat
44324409-5cb3-438a-9751-9ee868b35233
0.0.1
hfopenllm_v2/Felladrin_Minueza-32M-UltraChat/1762652579.6187
1762652579.6187022
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Felladrin/Minueza-32M-UltraChat
Felladrin/Minueza-32M-UltraChat
Felladrin
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.13756277787381924}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 0.033}
HF Open LLM v2
Felladrin
Felladrin/Llama-160M-Chat-v1
0885ef86-d7ef-4261-8ccd-f0391c42ffe4
0.0.1
hfopenllm_v2/Felladrin_Llama-160M-Chat-v1/1762652579.618279
1762652579.61828
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Felladrin/Llama-160M-Chat-v1
Felladrin/Llama-160M-Chat-v1
Felladrin
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15754642127333254}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 0.162}
HF Open LLM v2
kavonalds
kavonalds/BunderMaxx-1010
6b0275ea-f2eb-4a37-922c-d1f734c1a6d3
0.0.1
hfopenllm_v2/kavonalds_BunderMaxx-1010/1762652580.305197
1762652580.3051982
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
kavonalds/BunderMaxx-1010
kavonalds/BunderMaxx-1010
kavonalds
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2980558252104416}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.236}
HF Open LLM v2
kavonalds
kavonalds/BunderMaxx-0710
63d646bf-14d2-4cc7-ab82-efd1645cc1ba
0.0.1
hfopenllm_v2/kavonalds_BunderMaxx-0710/1762652580.3044312
1762652580.3044322
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
kavonalds/BunderMaxx-0710
kavonalds/BunderMaxx-0710
kavonalds
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.32825569488955975}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 1.236}
HF Open LLM v2
kavonalds
kavonalds/BunderMaxx-0710
10be7d08-18a9-43a6-80ea-81d704600eab
0.0.1
hfopenllm_v2/kavonalds_BunderMaxx-0710/1762652580.304877
1762652580.3048792
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
kavonalds/BunderMaxx-0710
kavonalds/BunderMaxx-0710
kavonalds
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.27007894608527594}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 1.236}
HF Open LLM v2
kavonalds
kavonalds/Lancer-1-1b-Instruct
ae2afa83-4607-43ea-be11-86cc57f3b848
0.0.1
hfopenllm_v2/kavonalds_Lancer-1-1b-Instruct/1762652580.305463
1762652580.305465
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
kavonalds/Lancer-1-1b-Instruct
kavonalds/Lancer-1-1b-Instruct
kavonalds
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5545940327220664}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 1.236}
HF Open LLM v2
olabs-ai
olabs-ai/reflection_model
84b63639-3343-4568-9fa7-d353ccb5b465
0.0.1
hfopenllm_v2/olabs-ai_reflection_model/1762652580.417324
1762652580.417325
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
olabs-ai/reflection_model
olabs-ai/reflection_model
olabs-ai
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15986914719610634}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "?", "params_billions": 9.3}
HF Open LLM v2
UCLA-AGI
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3
d8d05a10-8889-40aa-b56f-365e0a12052c
0.0.1
hfopenllm_v2/UCLA-AGI_Llama-3-Instruct-8B-SPPO-Iter3/1762652579.937166
1762652579.9371672
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3
UCLA-AGI
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6834122350917787}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
UCLA-AGI
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3
49e095af-ed90-4e64-b476-4fc62d6e6997
0.0.1
hfopenllm_v2/UCLA-AGI_Llama-3-Instruct-8B-SPPO-Iter3/1762652579.937367
1762652579.9373682
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3
UCLA-AGI
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.67029814226253}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH",...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
UCLA-AGI
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter2
0c5c315f-63c4-427e-a307-1422a197895c
0.0.1
hfopenllm_v2/UCLA-AGI_Llama-3-Instruct-8B-SPPO-Iter2/1762652579.93697
1762652579.936971
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter2
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter2
UCLA-AGI
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6988745417713889}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.0}
HF Open LLM v2
UCLA-AGI
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter1
07af3512-a045-435e-a965-8daa0836905d
0.0.1
hfopenllm_v2/UCLA-AGI_Llama-3-Instruct-8B-SPPO-Iter1/1762652579.9367309
1762652579.9367318
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter1
UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter1
UCLA-AGI
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7298988904994304}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
DeepMount00
DeepMount00/Lexora-Medium-7B
cc8f594a-e2f7-49e3-8654-57f1b397797f
0.0.1
hfopenllm_v2/DeepMount00_Lexora-Medium-7B/1762652579.551008
1762652579.551009
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DeepMount00/Lexora-Medium-7B
DeepMount00/Lexora-Medium-7B
DeepMount00
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4103379034295669}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
DeepMount00
DeepMount00/Lexora-Lite-3B_v2
bf38278f-6375-41a6-9744-04fb4a32ed72
0.0.1
hfopenllm_v2/DeepMount00_Lexora-Lite-3B_v2/1762652579.550789
1762652579.550789
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DeepMount00/Lexora-Lite-3B_v2
DeepMount00/Lexora-Lite-3B_v2
DeepMount00
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49431840848947456}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.086}
HF Open LLM v2
DeepMount00
DeepMount00/Lexora-Lite-3B
5eb28bbd-8428-4385-b078-13e8a868e9f0
0.0.1
hfopenllm_v2/DeepMount00_Lexora-Lite-3B/1762652579.550504
1762652579.550505
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DeepMount00/Lexora-Lite-3B
DeepMount00/Lexora-Lite-3B
DeepMount00
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5775996577968678}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 3.086}
HF Open LLM v2
DeepMount00
DeepMount00/Qwen2.5-7B-Instruct-MathCoder
ea1a36fb-66c0-4b1a-bdac-7ec2602a7c65
0.0.1
hfopenllm_v2/DeepMount00_Qwen2.5-7B-Instruct-MathCoder/1762652579.55323
1762652579.553231
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DeepMount00/Qwen2.5-7B-Instruct-MathCoder
DeepMount00/Qwen2.5-7B-Instruct-MathCoder
DeepMount00
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15302508455342934}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
DeepMount00
DeepMount00/mergekit-ties-okvgjfz
34350829-d42d-4e67-b23f-171044428c1f
0.0.1
hfopenllm_v2/DeepMount00_mergekit-ties-okvgjfz/1762652579.5535848
1762652579.553586
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DeepMount00/mergekit-ties-okvgjfz
DeepMount00/mergekit-ties-okvgjfz
DeepMount00
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15302508455342934}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 7.616}
HF Open LLM v2
PrimeIntellect
PrimeIntellect/INTELLECT-1-Instruct
ea823c15-3c92-4a67-a4fd-7826a9dd9e41
0.0.1
hfopenllm_v2/PrimeIntellect_INTELLECT-1-Instruct/1762652579.817848
1762652579.8178492
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
PrimeIntellect/INTELLECT-1-Instruct
PrimeIntellect/INTELLECT-1-Instruct
PrimeIntellect
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.0}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH", "lower_is_be...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.211}
HF Open LLM v2
PrimeIntellect
PrimeIntellect/INTELLECT-1
fee7966f-3e1b-43d9-b129-b0c23aac53b5
0.0.1
hfopenllm_v2/PrimeIntellect_INTELLECT-1/1762652579.817406
1762652579.817406
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
PrimeIntellect/INTELLECT-1
PrimeIntellect/INTELLECT-1
PrimeIntellect
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1757315035217667}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.211}
HF Open LLM v2
PrimeIntellect
PrimeIntellect/INTELLECT-1
bfffc240-22ab-4cc0-97c8-466ddf472ac4
0.0.1
hfopenllm_v2/PrimeIntellect_INTELLECT-1/1762652579.8176599
1762652579.817661
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
PrimeIntellect/INTELLECT-1
PrimeIntellect/INTELLECT-1
PrimeIntellect
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1757315035217667}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 10.211}
HF Open LLM v2
hotmailuser
hotmailuser/FalconSlerp-3B
2db7aa3c-4969-40c0-b8c6-1ff5c953ba23
0.0.1
hfopenllm_v2/hotmailuser_FalconSlerp-3B/1762652580.19236
1762652580.1923609
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/FalconSlerp-3B
hotmailuser/FalconSlerp-3B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5694568190179834}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 3.228}
HF Open LLM v2
hotmailuser
hotmailuser/Falcon3Slerp1-10B
376d342c-669b-4c76-9e7b-d49566ac441d
0.0.1
hfopenllm_v2/hotmailuser_Falcon3Slerp1-10B/1762652580.19171
1762652580.191711
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/Falcon3Slerp1-10B
hotmailuser/Falcon3Slerp1-10B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5694069513335727}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.306}
HF Open LLM v2
hotmailuser
hotmailuser/RombosBeagle-v2beta-MGS-32B
c507c0ac-759a-4013-8dd0-7ab5a959ca65
0.0.1
hfopenllm_v2/hotmailuser_RombosBeagle-v2beta-MGS-32B/1762652580.199307
1762652580.199308
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/RombosBeagle-v2beta-MGS-32B
hotmailuser/RombosBeagle-v2beta-MGS-32B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5156761836371937}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Qwen2ForCausalLM", "params_billions": 32.764}
HF Open LLM v2
hotmailuser
hotmailuser/FalconSlerp3-7B
017a681e-1bbb-4890-bfcc-f276954678e1
0.0.1
hfopenllm_v2/hotmailuser_FalconSlerp3-7B/1762652580.193249
1762652580.19325
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/FalconSlerp3-7B
hotmailuser/FalconSlerp3-7B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6096235765546527}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456}
HF Open LLM v2
hotmailuser
hotmailuser/Falcon3Slerp2-10B
bae0b772-8ae6-4fed-ae78-d6d83e560a95
0.0.1
hfopenllm_v2/hotmailuser_Falcon3Slerp2-10B/1762652580.191951
1762652580.191952
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/Falcon3Slerp2-10B
hotmailuser/Falcon3Slerp2-10B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6117966994241945}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.306}
HF Open LLM v2
hotmailuser
hotmailuser/Falcon3Slerp4-10B
d5466af4-2bef-4ce8-a659-9e05a5e674b6
0.0.1
hfopenllm_v2/hotmailuser_Falcon3Slerp4-10B/1762652580.19215
1762652580.192151
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/Falcon3Slerp4-10B
hotmailuser/Falcon3Slerp4-10B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6072254950198805}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 10.306}
HF Open LLM v2
hotmailuser
hotmailuser/FalconSlerp2-7B
fc8605ad-f7b9-4a73-afd3-85b996fc2549
0.0.1
hfopenllm_v2/hotmailuser_FalconSlerp2-7B/1762652580.1928341
1762652580.192835
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/FalconSlerp2-7B
hotmailuser/FalconSlerp2-7B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6160432097944565}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456}
HF Open LLM v2
hotmailuser
hotmailuser/FalconSlerp3-10B
f933fbc2-370e-4231-94a9-c833c2aa793d
0.0.1
hfopenllm_v2/hotmailuser_FalconSlerp3-10B/1762652580.1930392
1762652580.19304
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/FalconSlerp3-10B
hotmailuser/FalconSlerp3-10B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6001564737119731}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 10.306}
HF Open LLM v2
hotmailuser
hotmailuser/FalconSlerp1-7B
5d01fa6d-4280-4926-b166-e98892ee60f4
0.0.1
hfopenllm_v2/hotmailuser_FalconSlerp1-7B/1762652580.1925812
1762652580.192582
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/FalconSlerp1-7B
hotmailuser/FalconSlerp1-7B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5394564200765082}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456}
HF Open LLM v2
hotmailuser
hotmailuser/FalconSlerp6-7B
88a4587f-d3d4-4b08-b800-13a2daf4a660
0.0.1
hfopenllm_v2/hotmailuser_FalconSlerp6-7B/1762652580.193665
1762652580.193666
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/FalconSlerp6-7B
hotmailuser/FalconSlerp6-7B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6026542906155667}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456}
HF Open LLM v2
hotmailuser
hotmailuser/FalconSlerp4-7B
d6ac7c9f-212e-4000-b89e-d977122d2e2b
0.0.1
hfopenllm_v2/hotmailuser_FalconSlerp4-7B/1762652580.193457
1762652580.1934578
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/FalconSlerp4-7B
hotmailuser/FalconSlerp4-7B
hotmailuser
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6284580468711907}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.456}
HF Open LLM v2
Solshine
Solshine/Brimful-merged-replete
6523a08c-7a43-4784-9650-e1d5144fcfcf
0.0.1
hfopenllm_v2/Solshine_Brimful-merged-replete/1762652579.8890932
1762652579.8890939
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Solshine/Brimful-merged-replete
Solshine/Brimful-merged-replete
Solshine
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.17605619755581856}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "Qwen2ForCausalLM", "params_billions": 12.277}
HF Open LLM v2
oobabooga
oobabooga/CodeBooga-34B-v0.1
3086045f-e22d-4aca-9459-fc64454a2fb2
0.0.1
hfopenllm_v2/oobabooga_CodeBooga-34B-v0.1/1762652580.425838
1762652580.425838
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
oobabooga/CodeBooga-34B-v0.1
oobabooga/CodeBooga-34B-v0.1
oobabooga
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5250180631834643}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 33.744}
HF Open LLM v2
gaverfraxz
gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES
83a638be-6f3d-4d5b-b1de-6515634aebbd
0.0.1
hfopenllm_v2/gaverfraxz_Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES/1762652580.163549
1762652580.16355
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES
gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES
gaverfraxz
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45505148561372716}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
gaverfraxz
gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-DELLA
3666aa17-279d-4f0b-a6c2-2c8198729df9
0.0.1
hfopenllm_v2/gaverfraxz_Meta-Llama-3.1-8B-Instruct-HalfAbliterated-DELLA/1762652580.163272
1762652580.1632729
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-DELLA
gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-DELLA
gaverfraxz
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.40094615619888563}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
crestf411
crestf411/MN-Slush
b32a7808-7a64-41a8-aad4-030efc512906
0.0.1
hfopenllm_v2/crestf411_MN-Slush/1762652580.117737
1762652580.117738
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
crestf411/MN-Slush
crestf411/MN-Slush
crestf411
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4077148632295642}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
Invalid-Null
Invalid-Null/PeiYangMe-0.7
294c1745-38cb-4b1e-aae6-e2878ab9065a
0.0.1
hfopenllm_v2/Invalid-Null_PeiYangMe-0.7/1762652579.648521
1762652579.648522
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Invalid-Null/PeiYangMe-0.7
Invalid-Null/PeiYangMe-0.7
Invalid-Null
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1491032682172192}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 6.061}
HF Open LLM v2
Invalid-Null
Invalid-Null/PeiYangMe-0.5
c645a252-366a-4890-a16b-bf687bfbb593
0.0.1
hfopenllm_v2/Invalid-Null_PeiYangMe-0.5/1762652579.648252
1762652579.648252
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Invalid-Null/PeiYangMe-0.5
Invalid-Null/PeiYangMe-0.5
Invalid-Null
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.14088507382633633}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 6.061}
HF Open LLM v2
jaspionjader
jaspionjader/bh-28
0dc95982-e5b0-4011-9e5b-48af7e3048f0
0.0.1
hfopenllm_v2/jaspionjader_bh-28/1762652580.252297
1762652580.2522979
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-28
jaspionjader/bh-28
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4785070280189896}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-11
804f4be8-a8a9-473f-a898-d71b742a62eb
0.0.1
hfopenllm_v2/jaspionjader_bh-11/1762652580.2481
1762652580.2481012
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-11
jaspionjader/bh-11
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45754930251732073}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/test-13
98772920-a700-4fda-88fd-53c16ac4b1a1
0.0.1
hfopenllm_v2/jaspionjader_test-13/1762652580.278408
1762652580.278409
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/test-13
jaspionjader/test-13
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45809895521660304}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-v5-8B
53c89eb1-49ab-4e5f-b1ad-d8e80045a292
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-v5-8B/1762652580.2420359
1762652580.2420359
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-v5-8B
jaspionjader/Kosmos-EVAA-v5-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44595894448951}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH",...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/kstc-5-8b
ea79ca75-c55b-457a-b952-528a22567dbb
0.0.1
hfopenllm_v2/jaspionjader_kstc-5-8b/1762652580.270952
1762652580.270953
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/kstc-5-8b
jaspionjader/kstc-5-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47208739477918593}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-v4-8B
fd2a2a9c-639f-4348-9861-00271ed070b2
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-v4-8B/1762652580.241815
1762652580.241816
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-v4-8B
jaspionjader/Kosmos-EVAA-v4-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4289230353240513}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-8
f6dced28-f64c-4995-88b1-ac9a82903de2
0.0.1
hfopenllm_v2/jaspionjader_bh-8/1762652580.262149
1762652580.262152
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-8
jaspionjader/bh-8
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45967255770245175}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/f-4-8b
a98ec95c-4af0-4b55-adbc-06e5ceecd00f
0.0.1
hfopenllm_v2/jaspionjader_f-4-8b/1762652580.265391
1762652580.2653928
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/f-4-8b
jaspionjader/f-4-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4797060687863757}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Auro-Kosmos-EVAA-v2.2-8B
4e616fc6-8baa-4c9a-9098-b8d108911ad2
0.0.1
hfopenllm_v2/jaspionjader_Auro-Kosmos-EVAA-v2.2-8B/1762652580.231466
1762652580.231467
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Auro-Kosmos-EVAA-v2.2-8B
jaspionjader/Auro-Kosmos-EVAA-v2.2-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4267997801389203}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/test-16
80c756a7-9d47-4b49-bf42-bbada0909163
0.0.1
hfopenllm_v2/jaspionjader_test-16/1762652580.279189
1762652580.27919
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/test-16
jaspionjader/test-16
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4599473840520929}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015}
HF Open LLM v2
jaspionjader
jaspionjader/fr-10-8b
725e5a72-548f-46d0-b268-12209e5cb085
0.0.1
hfopenllm_v2/jaspionjader_fr-10-8b/1762652580.268136
1762652580.268136
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/fr-10-8b
jaspionjader/fr-10-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44018869931781013}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/f-3-8b
5ba1e4d3-29d4-4337-bd10-9e1a5df29af4
0.0.1
hfopenllm_v2/jaspionjader_f-3-8b/1762652580.264997
1762652580.264998
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/f-3-8b
jaspionjader/f-3-8b
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4803055891700687}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-8B
746ffa2c-cc95-4d69-9e46-0e8f4febd440
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-8B/1762652580.232065
1762652580.232065
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-8B
jaspionjader/Kosmos-EVAA-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4404635256674513}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-55
7c388cc5-fb2f-48ba-967c-a931fcb25a42
0.0.1
hfopenllm_v2/jaspionjader_bh-55/1762652580.259115
1762652580.259116
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-55
jaspionjader/bh-55
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47093822169621047}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-2
3e4b8dcc-9270-4b14-952f-c6b96ee8ce57
0.0.1
hfopenllm_v2/jaspionjader_bh-2/1762652580.250077
1762652580.250078
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-2
jaspionjader/bh-2
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45792386423578324}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-gamma-v14-8B
c20f5702-24fc-443a-875e-495401776eeb
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-v14-8B/1762652580.23884
1762652580.23884
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-gamma-v14-8B
jaspionjader/Kosmos-EVAA-gamma-v14-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4380155764482684}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/test-14
d647b482-3d3b-4ed4-b8b5-d57eedf87db9
0.0.1
hfopenllm_v2/jaspionjader_test-14/1762652580.2787268
1762652580.278728
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/test-14
jaspionjader/test-14
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4443853420036614}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015}
HF Open LLM v2
jaspionjader
jaspionjader/PRP-Kosmos-EVAA-8B
84a37d06-2668-4143-8e2f-5a08651f2dfb
0.0.1
hfopenllm_v2/jaspionjader_PRP-Kosmos-EVAA-8B/1762652580.244709
1762652580.24471
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/PRP-Kosmos-EVAA-8B
jaspionjader/PRP-Kosmos-EVAA-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.36327721556580983}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-10
86411dbb-e28b-4e9d-856e-fcc001252fbe
0.0.1
hfopenllm_v2/jaspionjader_bh-10/1762652580.247846
1762652580.2478468
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-10
jaspionjader/bh-10
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46184568057199343}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/bh-4
85ba493b-05f1-4853-a0ff-44570a7c2a82
0.0.1
hfopenllm_v2/jaspionjader_bh-4/1762652580.2554429
1762652580.255444
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/bh-4
jaspionjader/bh-4
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4672912317096415}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
jaspionjader
jaspionjader/Kosmos-EVAA-gamma-ultra-light-8B
1810feae-7a27-4c17-8174-3cd8a143b21f
0.0.1
hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-ultra-light-8B/1762652580.238316
1762652580.238317
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jaspionjader/Kosmos-EVAA-gamma-ultra-light-8B
jaspionjader/Kosmos-EVAA-gamma-ultra-light-8B
jaspionjader
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4563003940655239}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}