_leaderboard
stringclasses
1 value
_developer
stringclasses
559 values
_model
stringlengths
9
102
_uuid
stringlengths
36
36
schema_version
stringclasses
1 value
evaluation_id
stringlengths
35
133
retrieved_timestamp
stringlengths
13
18
source_data
stringclasses
1 value
evaluation_source_name
stringclasses
1 value
evaluation_source_type
stringclasses
1 value
source_organization_name
stringclasses
1 value
source_organization_url
null
source_organization_logo_url
null
evaluator_relationship
stringclasses
1 value
model_name
stringlengths
4
102
model_id
stringlengths
9
102
model_developer
stringclasses
559 values
model_inference_platform
stringclasses
1 value
evaluation_results
stringlengths
1.35k
1.41k
additional_details
stringclasses
660 values
HF Open LLM v2
google
Gunulhona/Gemma-Ko-Merge-PEFT
f9fb4008-db4e-4a84-b12b-050bdf35084f
0.0.1
hfopenllm_v2/Gunulhona_Gemma-Ko-Merge-PEFT/1762652579.635457
1762652579.635457
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Gunulhona/Gemma-Ko-Merge-PEFT
Gunulhona/Gemma-Ko-Merge-PEFT
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.28803906966847964}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "?", "params_billions": 20.318}
HF Open LLM v2
google
Gunulhona/Gemma-Ko-Merge-PEFT
7891a95c-8d95-4181-96e8-cdc2f6ab538b
0.0.1
hfopenllm_v2/Gunulhona_Gemma-Ko-Merge-PEFT/1762652579.635783
1762652579.635786
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Gunulhona/Gemma-Ko-Merge-PEFT
Gunulhona/Gemma-Ko-Merge-PEFT
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4441348954108433}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "?", "params_billions": 20.318}
HF Open LLM v2
google
ell44ot/gemma-2b-def
9ba31c7b-13df-46f2-a164-1729563707e1
0.0.1
hfopenllm_v2/ell44ot_gemma-2b-def/1762652580.147274
1762652580.147275
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ell44ot/gemma-2b-def
ell44ot/gemma-2b-def
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.26930433472076315}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "GemmaModel", "params_billions": 1.546}
HF Open LLM v2
google
BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference
0f948238-5ed2-41ee-a815-3ff20728de89
0.0.1
hfopenllm_v2/BAAI_Gemma2-9B-IT-Simpo-Infinity-Preference/1762652579.487571
1762652579.487571
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference
BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31763831079314}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH",...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
Triangle104/Gemmadevi-Stock-10B
153fd43a-fe54-4a99-98dd-5420f2bf8b66
0.0.1
hfopenllm_v2/Triangle104_Gemmadevi-Stock-10B/1762652579.9249291
1762652579.9249291
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Triangle104/Gemmadevi-Stock-10B
Triangle104/Gemmadevi-Stock-10B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15819470117067158}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
SaisExperiments/Gemma-2-2B-Stheno-Filtered
16070acb-e8bb-476a-b5aa-863a85cb0aee
0.0.1
hfopenllm_v2/SaisExperiments_Gemma-2-2B-Stheno-Filtered/1762652579.855671
1762652579.8556721
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
SaisExperiments/Gemma-2-2B-Stheno-Filtered
SaisExperiments/Gemma-2-2B-Stheno-Filtered
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4196554032190144}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
google/flan-t5-xl
ab0ac321-1c2b-4523-b48c-de47ff06e7a3
0.0.1
hfopenllm_v2/google_flan-t5-xl/1762652580.173602
1762652580.173603
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/flan-t5-xl
google/flan-t5-xl
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.22374189373085634}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "T5ForConditionalGeneration", "params_billions": 2.85}
HF Open LLM v2
google
google/flan-t5-xl
98a6a294-7b5d-4279-8aa6-6ed16248ce0b
0.0.1
hfopenllm_v2/google_flan-t5-xl/1762652580.1738272
1762652580.1738281
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/flan-t5-xl
google/flan-t5-xl
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2206944241279804}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "T5ForConditionalGeneration", "params_billions": 2.85}
HF Open LLM v2
google
cognitivecomputations/dolphin-2.9.4-gemma2-2b
29a10f53-dd38-437b-a7f3-9756035df640
0.0.1
hfopenllm_v2/cognitivecomputations_dolphin-2.9.4-gemma2-2b/1762652580.115823
1762652580.115823
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
cognitivecomputations/dolphin-2.9.4-gemma2-2b
cognitivecomputations/dolphin-2.9.4-gemma2-2b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.08955127949396491}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
agentlans/Gemma2-9B-AdvancedFuse
3bcdf1ca-ad29-45cf-ac97-6bc508981545
0.0.1
hfopenllm_v2/agentlans_Gemma2-9B-AdvancedFuse/1762652579.975734
1762652579.975735
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
agentlans/Gemma2-9B-AdvancedFuse
agentlans/Gemma2-9B-AdvancedFuse
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15427288483446144}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO-alpaca
3cc8621a-b38c-4735-af09-027989774289
0.0.1
hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-17-ORPO-alpaca/1762652580.6102881
1762652580.6102889
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO-alpaca
ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO-alpaca
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30647349033896726}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18-merge
9c7a213f-e5f8-4cc2-9cbe-d61db2cf2bbe
0.0.1
hfopenllm_v2/ymcki_gemma-2-2b-ORPO-jpn-it-abliterated-18-merge/1762652580.609323
1762652580.609324
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18-merge
ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18-merge
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5218209905273563}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
ymcki/gemma-2-2b-jpn-it-abliterated-24
4f0262d9-2a01-4127-bb40-1bbf437bbc07
0.0.1
hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-24/1762652580.610902
1762652580.610903
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ymcki/gemma-2-2b-jpn-it-abliterated-24
ymcki/gemma-2-2b-jpn-it-abliterated-24
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49786566310722213}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
ymcki/gemma-2-2b-jpn-it-abliterated-17-18-24
7321bd04-6f20-427a-8219-0ff2e299cb01
0.0.1
hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-17-18-24/1762652580.609858
1762652580.609859
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ymcki/gemma-2-2b-jpn-it-abliterated-17-18-24
ymcki/gemma-2-2b-jpn-it-abliterated-17-18-24
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.505484337114412}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
23800723-b5bd-4fc6-9d07-ca937c8680c6
0.0.1
hfopenllm_v2/ymcki_gemma-2-2b-ORPO-jpn-it-abliterated-18/1762652580.6090298
1762652580.609031
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4630945890237902}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
ymcki/gemma-2-2b-jpn-it-abliterated-18
78f235b0-fa98-48e2-bb03-9f7e9f986004
0.0.1
hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-18/1762652580.610494
1762652580.610495
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ymcki/gemma-2-2b-jpn-it-abliterated-18
ymcki/gemma-2-2b-jpn-it-abliterated-18
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5175246124726836}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO
44b47789-f529-4bae-9e87-196abc325efc
0.0.1
hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-17-ORPO/1762652580.610075
1762652580.610076
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO
ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47478468242042227}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
ymcki/gemma-2-2b-jpn-it-abliterated-18-ORPO
c91ab7d1-b36e-45ca-8f1e-ad9ef0c38100
0.0.1
hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-18-ORPO/1762652580.610698
1762652580.610699
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ymcki/gemma-2-2b-jpn-it-abliterated-18-ORPO
ymcki/gemma-2-2b-jpn-it-abliterated-18-ORPO
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47423502972113984}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
ymcki/gemma-2-2b-jpn-it-abliterated-17
5958a61d-bf39-4de4-bfe1-6a6db2f37f55
0.0.1
hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-17/1762652580.609628
1762652580.609628
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ymcki/gemma-2-2b-jpn-it-abliterated-17
ymcki/gemma-2-2b-jpn-it-abliterated-17
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5081572449988254}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
Youlln/4PRYMMAL-GEMMA2-9B-SLERP
06b75d54-4d17-4116-a4d5-0917eedb2dc4
0.0.1
hfopenllm_v2/Youlln_4PRYMMAL-GEMMA2-9B-SLERP/1762652579.961175
1762652579.9611762
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
Youlln/4PRYMMAL-GEMMA2-9B-SLERP
Youlln/4PRYMMAL-GEMMA2-9B-SLERP
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2713766140507188}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
google/flan-t5-large
eb2e1202-9292-4f5e-a366-abc84897c66d
0.0.1
hfopenllm_v2/google_flan-t5-large/1762652580.173132
1762652580.1731331
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/flan-t5-large
google/flan-t5-large
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.22009490374428736}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "T5ForConditionalGeneration", "params_billions": 0.783}
HF Open LLM v2
google
jebish7/gemma-2-2b-it
86206a02-3ab9-4a86-a00c-2900e8cd2e18
0.0.1
hfopenllm_v2/jebish7_gemma-2-2b-it/1762652580.2824588
1762652580.2824588
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jebish7/gemma-2-2b-it
jebish7/gemma-2-2b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.12717035244263}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH",...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
jebish7/gemma-2-9b-it
80a35d79-893b-439f-b100-a538a3c86974
0.0.1
hfopenllm_v2/jebish7_gemma-2-9b-it/1762652580.282719
1762652580.28272
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
jebish7/gemma-2-9b-it
jebish7/gemma-2-9b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1557467519514887}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
google/mt5-xxl
38520cce-b3b6-4f22-a6a8-313f6181f5ea
0.0.1
hfopenllm_v2/google_mt5-xxl/1762652580.1791801
1762652580.1791801
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/mt5-xxl
google/mt5-xxl
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.23575668116154028}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "T5ForConditionalGeneration", "params_billions": 11.9}
HF Open LLM v2
google
princeton-nlp/gemma-2-9b-it-SimPO
4285b38c-aba8-444b-9b0b-b265c7b1fef1
0.0.1
hfopenllm_v2/princeton-nlp_gemma-2-9b-it-SimPO/1762652580.454763
1762652580.4547682
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
princeton-nlp/gemma-2-9b-it-SimPO
princeton-nlp/gemma-2-9b-it-SimPO
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3206857803960159}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
VAGOsolutions/SauerkrautLM-Gemma-2b
b002a274-9b4f-40ad-b0c7-e4efabbe431f
0.0.1
hfopenllm_v2/VAGOsolutions_SauerkrautLM-Gemma-2b/1762652579.941349
1762652579.94135
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
VAGOsolutions/SauerkrautLM-Gemma-2b
VAGOsolutions/SauerkrautLM-Gemma-2b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24752213017017072}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 2.506}
HF Open LLM v2
google
VAGOsolutions/SauerkrautLM-gemma-2-2b-it
b010858c-edb5-4e49-b5b6-72b06943ab2c
0.0.1
hfopenllm_v2/VAGOsolutions_SauerkrautLM-gemma-2-2b-it/1762652579.9427688
1762652579.94277
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
VAGOsolutions/SauerkrautLM-gemma-2-2b-it
VAGOsolutions/SauerkrautLM-gemma-2-2b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.13206625088099574}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
VAGOsolutions/SauerkrautLM-gemma-2-9b-it
5395cbac-afe0-4936-b4eb-f554fcb5be75
0.0.1
hfopenllm_v2/VAGOsolutions_SauerkrautLM-gemma-2-9b-it/1762652579.94298
1762652579.942981
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
VAGOsolutions/SauerkrautLM-gemma-2-9b-it
VAGOsolutions/SauerkrautLM-gemma-2-9b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3024009627787604}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
VAGOsolutions/SauerkrautLM-Gemma-7b
e66f4326-2585-4581-b45f-d9a81fb1576c
0.0.1
hfopenllm_v2/VAGOsolutions_SauerkrautLM-Gemma-7b/1762652579.9415941
1762652579.9415948
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
VAGOsolutions/SauerkrautLM-Gemma-7b
VAGOsolutions/SauerkrautLM-Gemma-7b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3406705319662939}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 8.538}
HF Open LLM v2
google
google/mt5-small
0d958c7c-5cd9-459f-a0e9-235b5d41ae53
0.0.1
hfopenllm_v2/google_mt5-small/1762652580.1787279
1762652580.178729
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/mt5-small
google/mt5-small
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.17180968718555653}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "MT5ForConditionalGeneration", "params_billions": 0.17}
HF Open LLM v2
google
ZHLiu627/zephyr-7b-gemma-rpo-avg
6333359d-1cf7-4905-9a48-f8a8f7b46ed2
0.0.1
hfopenllm_v2/ZHLiu627_zephyr-7b-gemma-rpo-avg/1762652579.9660559
1762652579.966057
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ZHLiu627/zephyr-7b-gemma-rpo-avg
ZHLiu627/zephyr-7b-gemma-rpo-avg
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30060350979844586}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "GemmaForCausalLM", "params_billions": 8.538}
HF Open LLM v2
google
anthracite-org/magnum-v3-9b-customgemma2
865b86aa-7b8d-4619-aa57-3c57cc4c7b51
0.0.1
hfopenllm_v2/anthracite-org_magnum-v3-9b-customgemma2/1762652580.012768
1762652580.012769
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
anthracite-org/magnum-v3-9b-customgemma2
anthracite-org/magnum-v3-9b-customgemma2
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1272955757390391}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
qq8933/OpenLongCoT-Base-Gemma2-2B
c945b9b5-7b46-4300-adcc-2d6c94df0ac1
0.0.1
hfopenllm_v2/qq8933_OpenLongCoT-Base-Gemma2-2B/1762652580.488883
1762652580.488883
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
qq8933/OpenLongCoT-Base-Gemma2-2B
qq8933/OpenLongCoT-Base-Gemma2-2B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1965141380426158}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 3.204}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v2-9B
ca1b9625-0112-4ebf-b1c3-d2dd217d50b2
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v2-9B/1762652580.315539
1762652580.31554
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v2-9B
lemon07r/Gemma-2-Ataraxy-v2-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.21362429464930827}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v4c-9B
9499ec24-5be2-478c-b13e-3102d1555668
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v4c-9B/1762652580.318075
1762652580.318076
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v4c-9B
lemon07r/Gemma-2-Ataraxy-v4c-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6945282960323054}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v4a-Advanced-9B
b84aedba-7b87-445d-87c2-b029cb0038c3
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v4a-Advanced-9B/1762652580.317515
1762652580.317516
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v4a-Advanced-9B
lemon07r/Gemma-2-Ataraxy-v4a-Advanced-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7135123694020753}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v3i-9B
53602c70-73d9-461b-b27a-24c6a1a538e5
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v3i-9B/1762652580.3165948
1762652580.316596
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v3i-9B
lemon07r/Gemma-2-Ataraxy-v3i-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4203047912871182}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v3j-9B
d435bd27-1c26-429d-8ac5-8fd8c591a9aa
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v3j-9B/1762652580.3168168
1762652580.316818
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v3j-9B
lemon07r/Gemma-2-Ataraxy-v3j-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4169326276501904}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v4-Advanced-9B
c0e95e3f-37a4-4b2f-a37b-37854546c241
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v4-Advanced-9B/1762652580.317157
1762652580.3171608
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v4-Advanced-9B
lemon07r/Gemma-2-Ataraxy-v4-Advanced-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7015474496558022}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-9B
9ba72d50-4321-4383-8be9-286a56607624
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-9B/1762652580.31483
1762652580.314831
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-9B
lemon07r/Gemma-2-Ataraxy-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3008772279773224}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v2a-9B
4fa1e172-f570-4a96-b53a-8ecf31854191
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v2a-9B/1762652580.315754
1762652580.315755
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v2a-9B
lemon07r/Gemma-2-Ataraxy-v2a-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15946909755005606}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v2f-9B
fd59fb1c-3681-44d2-9172-b10891ae9c55
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v2f-9B/1762652580.315967
1762652580.315968
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v2f-9B
lemon07r/Gemma-2-Ataraxy-v2f-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.37911408396388246}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v3-Advanced-9B
778a10b0-c537-4592-9dbb-2b0de07ced4c
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v3-Advanced-9B/1762652580.316169
1762652580.316169
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v3-Advanced-9B
lemon07r/Gemma-2-Ataraxy-v3-Advanced-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6601816513517467}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v4b-9B
41f04f45-2f1d-42fd-87de-cc5e484cada2
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v4b-9B/1762652580.317803
1762652580.317804
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v4b-9B
lemon07r/Gemma-2-Ataraxy-v4b-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6878338364428604}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-Remix-9B
29dfbb00-8760-46d8-bef8-d036870fb0c0
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-Remix-9B/1762652580.31531
1762652580.3153112
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-Remix-9B
lemon07r/Gemma-2-Ataraxy-Remix-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7083416446140685}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-Advanced-9B
7806d1aa-b9e2-45bc-b89d-76e6c48dd3a0
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-Advanced-9B/1762652580.315091
1762652580.315092
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-Advanced-9B
lemon07r/Gemma-2-Ataraxy-Advanced-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5515964308036011}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v4d-9B
7e6685d8-af21-4810-a9cc-edb296f4b937
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v4d-9B/1762652580.318495
1762652580.318496
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v4d-9B
lemon07r/Gemma-2-Ataraxy-v4d-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7250029920610646}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
lemon07r/Gemma-2-Ataraxy-v3b-9B
d048e6ad-cc57-4ebe-8376-262564e86f0c
0.0.1
hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v3b-9B/1762652580.3163798
1762652580.316381
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
lemon07r/Gemma-2-Ataraxy-v3b-9B
lemon07r/Gemma-2-Ataraxy-v3b-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6809144181881852}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
ModelSpace/GemmaX2-28-9B-v0.1
6cb560eb-08f5-4430-8797-1116f1d2f56c
0.0.1
hfopenllm_v2/ModelSpace_GemmaX2-28-9B-v0.1/1762652579.76179
1762652579.761791
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
ModelSpace/GemmaX2-28-9B-v0.1
ModelSpace/GemmaX2-28-9B-v0.1
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.003921816336210145}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on B...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
google/gemma-2-27b
12f7d5a6-3f8b-49d8-9ca8-38774dbcca92
0.0.1
hfopenllm_v2/google_gemma-2-27b/1762652580.175144
1762652580.175145
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2-27b
google/gemma-2-27b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24752213017017072}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
google/gemma-2-2b
53fb75b1-2d9f-4af3-a358-18bf5d4a9032
0.0.1
hfopenllm_v2/google_gemma-2-2b/1762652580.1759539
1762652580.175955
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2-2b
google/gemma-2-2b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20176021844262113}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "InternLM2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
google/gemma-2-2b
07e74f27-e0c3-448f-9a8c-a07ff8a73178
0.0.1
hfopenllm_v2/google_gemma-2-2b/1762652580.175597
1762652580.1755981
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2-2b
google/gemma-2-2b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.19931226922343825}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "InternLM2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
google/gemma-2-2b-it
64daa9ea-cf1e-4787-90cf-ed72c5e23afd
0.0.1
hfopenllm_v2/google_gemma-2-2b-it/1762652580.176172
1762652580.176194
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2-2b-it
google/gemma-2-2b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5668337788179807}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "InternLM2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
google/recurrentgemma-2b
218a5d0f-5242-43c4-8166-81f5c09626bb
0.0.1
hfopenllm_v2/google_recurrentgemma-2b/1762652580.179393
1762652580.179394
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/recurrentgemma-2b
google/recurrentgemma-2b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3017028151970106}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "RecurrentGemmaForCausalLM", "params_billions": 2.683}
HF Open LLM v2
google
google/recurrentgemma-2b-it
a219b160-3dbd-4dcd-b39d-d12c6f9b1145
0.0.1
hfopenllm_v2/google_recurrentgemma-2b-it/1762652580.17961
1762652580.179611
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/recurrentgemma-2b-it
google/recurrentgemma-2b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2949329999955673}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "RecurrentGemmaForCausalLM", "params_billions": 2.683}
HF Open LLM v2
google
google/codegemma-1.1-2b
9d92e421-c458-4ad3-b9bf-45c0ca1b90cf
0.0.1
hfopenllm_v2/google_codegemma-1.1-2b/1762652580.172607
1762652580.172608
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/codegemma-1.1-2b
google/codegemma-1.1-2b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.22936253584932426}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 2.506}
HF Open LLM v2
google
google/gemma-1.1-7b-it
6929c338-76a5-4386-9fa8-68e35a989a86
0.0.1
hfopenllm_v2/google_gemma-1.1-7b-it/1762652580.1748302
1762652580.1748302
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-1.1-7b-it
google/gemma-1.1-7b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5039107346285633}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 8.538}
HF Open LLM v2
google
google/gemma-2-27b-it
5bcf96ce-efd1-4f90-91a1-edd548de71ad
0.0.1
hfopenllm_v2/google_gemma-2-27b-it/1762652580.17537
1762652580.175371
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2-27b-it
google/gemma-2-27b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7977677008116243}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
google/gemma-2-9b-it
e8cef406-d6cc-48bd-872f-3d5b74bcf092
0.0.1
hfopenllm_v2/google_gemma-2-9b-it/1762652580.177257
1762652580.177258
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2-9b-it
google/gemma-2-9b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7435626360279614}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.0}
HF Open LLM v2
google
google/gemma-2b-it
50dffd1a-ddf5-40fd-a2c8-e5dd140af617
0.0.1
hfopenllm_v2/google_gemma-2b-it/1762652580.17777
1762652580.17777
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2b-it
google/gemma-2b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.26902950837112194}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 2.506}
HF Open LLM v2
google
google/recurrentgemma-9b
1ff3ab95-3007-4cbf-a146-5e8e4ae65404
0.0.1
hfopenllm_v2/google_recurrentgemma-9b/1762652580.17984
1762652580.179841
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/recurrentgemma-9b
google/recurrentgemma-9b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31159434744256354}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "RecurrentGemmaForCausalLM", "params_billions": 9.0}
HF Open LLM v2
google
google/gemma-7b
630e3cc0-fccc-41b3-b439-85a875dae401
0.0.1
hfopenllm_v2/google_gemma-7b/1762652580.1780128
1762652580.178014
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-7b
google/gemma-7b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2659321710838353}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 8.538}
HF Open LLM v2
google
google/gemma-2-9b
2ac50111-a850-4bd2-8136-c373990742a5
0.0.1
hfopenllm_v2/google_gemma-2-9b/1762652580.177011
1762652580.177012
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2-9b
google/gemma-2-9b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20398320899657355}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.0}
HF Open LLM v2
google
google/gemma-2b
2dd86ebc-0253-4801-ac99-2bb3494ad29b
0.0.1
hfopenllm_v2/google_gemma-2b/1762652580.177512
1762652580.177513
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2b
google/gemma-2b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20375825033134307}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 2.506}
HF Open LLM v2
google
google/gemma-7b-it
30146048-ee0f-431d-b3e7-8c066c820740
0.0.1
hfopenllm_v2/google_gemma-7b-it/1762652580.178242
1762652580.1782432
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-7b-it
google/gemma-7b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3868324933398937}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 8.538}
HF Open LLM v2
google
google/gemma-1.1-2b-it
5ed676b6-4aff-4d71-a91a-6d5d9feeb28f
0.0.1
hfopenllm_v2/google_gemma-1.1-2b-it/1762652580.1745641
1762652580.174565
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-1.1-2b-it
google/gemma-1.1-2b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30674831668860847}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 2.506}
HF Open LLM v2
google
google/recurrentgemma-9b-it
c7095b76-2d50-467b-a8d9-d7a277f1f14c
0.0.1
hfopenllm_v2/google_recurrentgemma-9b-it/1762652580.180049
1762652580.18005
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/recurrentgemma-9b-it
google/recurrentgemma-9b-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5010383560065071}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "RecurrentGemmaForCausalLM", "params_billions": 9.0}
HF Open LLM v2
google
google/gemma-2-2b-jpn-it
251b93fa-6f12-41bc-85c8-ded52e1a0d2d
0.0.1
hfopenllm_v2/google_gemma-2-2b-jpn-it/1762652580.1767948
1762652580.176796
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2-2b-jpn-it
google/gemma-2-2b-jpn-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5288401441508531}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
google/gemma-2-2b-jpn-it
a09fdbce-489c-4d14-a05f-7663121bece7
0.0.1
hfopenllm_v2/google_gemma-2-2b-jpn-it/1762652580.176506
1762652580.176507
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/gemma-2-2b-jpn-it
google/gemma-2-2b-jpn-it
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5077826832803628}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
google/flan-ul2
99941572-3e23-467c-97df-dfe1a2aa9805
0.0.1
hfopenllm_v2/google_flan-ul2/1762652580.174251
1762652580.174251
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/flan-ul2
google/flan-ul2
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.23925406809487715}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "T5ForConditionalGeneration", "params_billions": 19.46}
HF Open LLM v2
google
wzhouad/gemma-2-9b-it-WPO-HB
70fe199f-6c81-4d99-a595-208b7abc321f
0.0.1
hfopenllm_v2/wzhouad_gemma-2-9b-it-WPO-HB/1762652580.596365
1762652580.5963662
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
wzhouad/gemma-2-9b-it-WPO-HB
wzhouad/gemma-2-9b-it-WPO-HB
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5437029304467702}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
UCLA-AGI/Gemma-2-9B-It-SPPO-Iter2
fa584f01-69eb-4ecc-9f0d-049b6bfb05c8
0.0.1
hfopenllm_v2/UCLA-AGI_Gemma-2-9B-It-SPPO-Iter2/1762652579.936279
1762652579.93628
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
UCLA-AGI/Gemma-2-9B-It-SPPO-Iter2
UCLA-AGI/Gemma-2-9B-It-SPPO-Iter2
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3100196367859502}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3
f318d457-d295-4447-9222-0b0d92708b5d
0.0.1
hfopenllm_v2/UCLA-AGI_Gemma-2-9B-It-SPPO-Iter3/1762652579.9364889
1762652579.93649
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3
UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31671409637539505}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
UCLA-AGI/Gemma-2-9B-It-SPPO-Iter1
687769ed-44e9-4f3d-aee6-2dc4e98dd7ee
0.0.1
hfopenllm_v2/UCLA-AGI_Gemma-2-9B-It-SPPO-Iter1/1762652579.936019
1762652579.93602
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
UCLA-AGI/Gemma-2-9B-It-SPPO-Iter1
UCLA-AGI/Gemma-2-9B-It-SPPO-Iter1
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.308221075634871}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
hotmailuser/Gemma2Crono-27B
501e2a2c-e32c-455e-8e5f-f8bde053fddc
0.0.1
hfopenllm_v2/hotmailuser_Gemma2Crono-27B/1762652580.193866
1762652580.193866
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/Gemma2Crono-27B
hotmailuser/Gemma2Crono-27B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7086164709637096}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
hotmailuser/Gemma2atlas-27B
c9020f27-9175-4f12-a108-6cbff1c0cb22
0.0.1
hfopenllm_v2/hotmailuser_Gemma2atlas-27B/1762652580.1943119
1762652580.194313
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/Gemma2atlas-27B
hotmailuser/Gemma2atlas-27B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7213560020744957}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
hotmailuser/Gemma2magnum-27b
0ad192a1-b33f-4362-a21d-ccc590986c5c
0.0.1
hfopenllm_v2/hotmailuser_Gemma2magnum-27b/1762652580.1945128
1762652580.194514
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/Gemma2magnum-27b
hotmailuser/Gemma2magnum-27b
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5050599077115387}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
hotmailuser/Gemma2SimPO-27B
433a8abf-8ff7-40bb-a4d0-654efdb6bf86
0.0.1
hfopenllm_v2/hotmailuser_Gemma2SimPO-27B/1762652580.194106
1762652580.1941068
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
hotmailuser/Gemma2SimPO-27B
hotmailuser/Gemma2SimPO-27B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7222303488078299}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
sequelbox/gemma-2-9B-MOTH
4bdefb85-2413-43b7-8938-869ad0cff58f
0.0.1
hfopenllm_v2/sequelbox_gemma-2-9B-MOTH/1762652580.5126731
1762652580.512674
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
sequelbox/gemma-2-9B-MOTH
sequelbox/gemma-2-9B-MOTH
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20588150551647405}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
recoilme/recoilme-gemma-2-9B-v0.5
8fe5a1e8-1491-4e64-8aed-32e73f2dae6e
0.0.1
hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.5/1762652580.4931269
1762652580.493134
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
recoilme/recoilme-gemma-2-9B-v0.5
recoilme/recoilme-gemma-2-9B-v0.5
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7664186580495308}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
recoilme/recoilme-gemma-2-9B-v0.3
8d3bd687-89f5-4d62-af46-93646aea4341
0.0.1
hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.3/1762652580.492666
1762652580.492667
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
recoilme/recoilme-gemma-2-9B-v0.3
recoilme/recoilme-gemma-2-9B-v0.3
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.57607592299543}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH",...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
recoilme/recoilme-gemma-2-9B-v0.3
47cfe707-ba31-4c9b-aa15-9ab8b566e206
0.0.1
hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.3/1762652580.492416
1762652580.492416
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
recoilme/recoilme-gemma-2-9B-v0.3
recoilme/recoilme-gemma-2-9B-v0.3
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.743937197746424}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
recoilme/recoilme-gemma-2-9B-v0.2
6a15378c-36cc-4f5e-b184-5a19a6fbb192
0.0.1
hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.2/1762652580.492019
1762652580.49202
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
recoilme/recoilme-gemma-2-9B-v0.2
recoilme/recoilme-gemma-2-9B-v0.2
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7591745457608035}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
recoilme/recoilme-gemma-2-9B-v0.2
5826c93f-3642-44cf-b385-4a5ab5103086
0.0.1
hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.2/1762652580.4922318
1762652580.492233
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
recoilme/recoilme-gemma-2-9B-v0.2
recoilme/recoilme-gemma-2-9B-v0.2
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2746989100032359}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
recoilme/recoilme-gemma-2-9B-v0.4
28eef1b7-a83e-49c9-8f11-ef9e4ae7e1ce
0.0.1
hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.4/1762652580.4928808
1762652580.492882
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
recoilme/recoilme-gemma-2-9B-v0.4
recoilme/recoilme-gemma-2-9B-v0.4
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2561891337207498}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp
0a685d8f-38c7-4521-9613-7b36ad1cac73
0.0.1
hfopenllm_v2/recoilme_Gemma-2-Ataraxy-Gemmasutra-9B-slerp/1762652580.491603
1762652580.491603
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp
recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.28536505361330156}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp
054a662a-e425-448c-9556-6998833e51ff
0.0.1
hfopenllm_v2/recoilme_Gemma-2-Ataraxy-Gemmasutra-9B-slerp/1762652580.491333
1762652580.491333
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp
recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7648949232480928}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
recoilme/recoilme-gemma-2-9B-v0.1
d31a41b0-6500-4e1b-8435-b9d3e9725c02
0.0.1
hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.1/1762652580.491797
1762652580.491798
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
recoilme/recoilme-gemma-2-9B-v0.1
recoilme/recoilme-gemma-2-9B-v0.1
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.751506004069203}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
allknowingroger/Gemma2Slerp2-27B
1f2c33e8-2d7b-4bd5-81e8-1c9bcae0ae8f
0.0.1
hfopenllm_v2/allknowingroger_Gemma2Slerp2-27B/1762652579.986531
1762652579.9865322
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/Gemma2Slerp2-27B
allknowingroger/Gemma2Slerp2-27B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7545534736720789}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
allknowingroger/Gemma2Slerp2-2.6B
eeb46285-0c8d-43b7-9b6d-e86c24064fde
0.0.1
hfopenllm_v2/allknowingroger_Gemma2Slerp2-2.6B/1762652579.98633
1762652579.98633
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/Gemma2Slerp2-2.6B
allknowingroger/Gemma2Slerp2-2.6B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5747272791748117}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
allknowingroger/Gemma2Slerp1-27B
42d79295-bdb0-411d-b1b0-5cff954e925c
0.0.1
hfopenllm_v2/allknowingroger_Gemma2Slerp1-27B/1762652579.986121
1762652579.986122
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/Gemma2Slerp1-27B
allknowingroger/Gemma2Slerp1-27B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7186332265056716}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
allknowingroger/GemmaSlerp4-10B
32e38c82-d412-4888-9d9d-f89aef0989fd
0.0.1
hfopenllm_v2/allknowingroger_GemmaSlerp4-10B/1762652579.9875991
1762652579.9875998
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/GemmaSlerp4-10B
allknowingroger/GemmaSlerp4-10B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7326216660682544}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
allknowingroger/GemmaSlerp-9B
3aed9fd2-45bd-4568-8885-7fc2370bb26d
0.0.1
hfopenllm_v2/allknowingroger_GemmaSlerp-9B/1762652579.987181
1762652579.9871821
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/GemmaSlerp-9B
allknowingroger/GemmaSlerp-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.704320092909037}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
allknowingroger/GemmaSlerp5-10B
e325b56f-4306-4e37-adc5-c09b300a8c30
0.0.1
hfopenllm_v2/allknowingroger_GemmaSlerp5-10B/1762652579.9878101
1762652579.987811
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/GemmaSlerp5-10B
allknowingroger/GemmaSlerp5-10B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7353444416370785}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159}
HF Open LLM v2
google
allknowingroger/GemmaSlerp2-9B
99333370-c7d5-4763-b3a4-14adde0fab9e
0.0.1
hfopenllm_v2/allknowingroger_GemmaSlerp2-9B/1762652579.987394
1762652579.987395
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/GemmaSlerp2-9B
allknowingroger/GemmaSlerp2-9B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7281003293483512}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242}
HF Open LLM v2
google
allknowingroger/Gemma2Slerp3-27B
648810d4-4dd5-48c7-a4d7-b3d9d2f3f3f2
0.0.1
hfopenllm_v2/allknowingroger_Gemma2Slerp3-27B/1762652579.986752
1762652579.986753
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/Gemma2Slerp3-27B
allknowingroger/Gemma2Slerp3-27B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7426384216102164}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
allknowingroger/Gemma2Slerp1-2.6B
e52ac657-26a3-499a-949f-bf2a0b620d8e
0.0.1
hfopenllm_v2/allknowingroger_Gemma2Slerp1-2.6B/1762652579.985875
1762652579.985876
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/Gemma2Slerp1-2.6B
allknowingroger/Gemma2Slerp1-2.6B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5354348683714766}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614}
HF Open LLM v2
google
allknowingroger/Gemma2Slerp4-27B
f94f3bf1-cf85-4673-a5cf-368f250233e4
0.0.1
hfopenllm_v2/allknowingroger_Gemma2Slerp4-27B/1762652579.986965
1762652579.9869661
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/Gemma2Slerp4-27B
allknowingroger/Gemma2Slerp4-27B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7496575752337131}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
allknowingroger/GemmaStock1-27B
0b19d8bb-1952-4515-8d29-e55e1106e92b
0.0.1
hfopenllm_v2/allknowingroger_GemmaStock1-27B/1762652579.9880252
1762652579.9880252
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
allknowingroger/GemmaStock1-27B
allknowingroger/GemmaStock1-27B
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7509064836855099}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227}
HF Open LLM v2
google
google/mt5-xl
5abb3ce9-6ad4-4dfa-8bca-81ec6cb84426
0.0.1
hfopenllm_v2/google_mt5-xl/1762652580.17897
1762652580.1789708
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
google/mt5-xl
google/mt5-xl
google
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.19596448534333347}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "float16", "architecture": "MT5ForConditionalGeneration", "params_billions": 3.23}