_leaderboard stringclasses 1
value | _developer stringclasses 559
values | _model stringlengths 9 102 | _uuid stringlengths 36 36 | schema_version stringclasses 1
value | evaluation_id stringlengths 35 133 | retrieved_timestamp stringlengths 13 18 | source_data stringclasses 1
value | evaluation_source_name stringclasses 1
value | evaluation_source_type stringclasses 1
value | source_organization_name stringclasses 1
value | source_organization_url null | source_organization_logo_url null | evaluator_relationship stringclasses 1
value | model_name stringlengths 4 102 | model_id stringlengths 9 102 | model_developer stringclasses 559
values | model_inference_platform stringclasses 1
value | evaluation_results stringlengths 1.35k 1.41k | additional_details stringclasses 660
values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
HF Open LLM v2 | google | Gunulhona/Gemma-Ko-Merge-PEFT | f9fb4008-db4e-4a84-b12b-050bdf35084f | 0.0.1 | hfopenllm_v2/Gunulhona_Gemma-Ko-Merge-PEFT/1762652579.635457 | 1762652579.635457 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Gunulhona/Gemma-Ko-Merge-PEFT | Gunulhona/Gemma-Ko-Merge-PEFT | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.28803906966847964}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "?", "params_billions": 20.318} |
HF Open LLM v2 | google | Gunulhona/Gemma-Ko-Merge-PEFT | 7891a95c-8d95-4181-96e8-cdc2f6ab538b | 0.0.1 | hfopenllm_v2/Gunulhona_Gemma-Ko-Merge-PEFT/1762652579.635783 | 1762652579.635786 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Gunulhona/Gemma-Ko-Merge-PEFT | Gunulhona/Gemma-Ko-Merge-PEFT | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4441348954108433}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "?", "params_billions": 20.318} |
HF Open LLM v2 | google | ell44ot/gemma-2b-def | 9ba31c7b-13df-46f2-a164-1729563707e1 | 0.0.1 | hfopenllm_v2/ell44ot_gemma-2b-def/1762652580.147274 | 1762652580.147275 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ell44ot/gemma-2b-def | ell44ot/gemma-2b-def | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.26930433472076315}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "GemmaModel", "params_billions": 1.546} |
HF Open LLM v2 | google | BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference | 0f948238-5ed2-41ee-a815-3ff20728de89 | 0.0.1 | hfopenllm_v2/BAAI_Gemma2-9B-IT-Simpo-Infinity-Preference/1762652579.487571 | 1762652579.487571 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference | BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31763831079314}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH",... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | Triangle104/Gemmadevi-Stock-10B | 153fd43a-fe54-4a99-98dd-5420f2bf8b66 | 0.0.1 | hfopenllm_v2/Triangle104_Gemmadevi-Stock-10B/1762652579.9249291 | 1762652579.9249291 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Triangle104/Gemmadevi-Stock-10B | Triangle104/Gemmadevi-Stock-10B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15819470117067158}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | SaisExperiments/Gemma-2-2B-Stheno-Filtered | 16070acb-e8bb-476a-b5aa-863a85cb0aee | 0.0.1 | hfopenllm_v2/SaisExperiments_Gemma-2-2B-Stheno-Filtered/1762652579.855671 | 1762652579.8556721 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | SaisExperiments/Gemma-2-2B-Stheno-Filtered | SaisExperiments/Gemma-2-2B-Stheno-Filtered | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4196554032190144}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | google/flan-t5-xl | ab0ac321-1c2b-4523-b48c-de47ff06e7a3 | 0.0.1 | hfopenllm_v2/google_flan-t5-xl/1762652580.173602 | 1762652580.173603 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/flan-t5-xl | google/flan-t5-xl | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.22374189373085634}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "T5ForConditionalGeneration", "params_billions": 2.85} |
HF Open LLM v2 | google | google/flan-t5-xl | 98a6a294-7b5d-4279-8aa6-6ed16248ce0b | 0.0.1 | hfopenllm_v2/google_flan-t5-xl/1762652580.1738272 | 1762652580.1738281 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/flan-t5-xl | google/flan-t5-xl | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2206944241279804}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "T5ForConditionalGeneration", "params_billions": 2.85} |
HF Open LLM v2 | google | cognitivecomputations/dolphin-2.9.4-gemma2-2b | 29a10f53-dd38-437b-a7f3-9756035df640 | 0.0.1 | hfopenllm_v2/cognitivecomputations_dolphin-2.9.4-gemma2-2b/1762652580.115823 | 1762652580.115823 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | cognitivecomputations/dolphin-2.9.4-gemma2-2b | cognitivecomputations/dolphin-2.9.4-gemma2-2b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.08955127949396491}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | agentlans/Gemma2-9B-AdvancedFuse | 3bcdf1ca-ad29-45cf-ac97-6bc508981545 | 0.0.1 | hfopenllm_v2/agentlans_Gemma2-9B-AdvancedFuse/1762652579.975734 | 1762652579.975735 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | agentlans/Gemma2-9B-AdvancedFuse | agentlans/Gemma2-9B-AdvancedFuse | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15427288483446144}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO-alpaca | 3cc8621a-b38c-4735-af09-027989774289 | 0.0.1 | hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-17-ORPO-alpaca/1762652580.6102881 | 1762652580.6102889 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO-alpaca | ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO-alpaca | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30647349033896726}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18-merge | 9c7a213f-e5f8-4cc2-9cbe-d61db2cf2bbe | 0.0.1 | hfopenllm_v2/ymcki_gemma-2-2b-ORPO-jpn-it-abliterated-18-merge/1762652580.609323 | 1762652580.609324 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18-merge | ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18-merge | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5218209905273563}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | ymcki/gemma-2-2b-jpn-it-abliterated-24 | 4f0262d9-2a01-4127-bb40-1bbf437bbc07 | 0.0.1 | hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-24/1762652580.610902 | 1762652580.610903 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ymcki/gemma-2-2b-jpn-it-abliterated-24 | ymcki/gemma-2-2b-jpn-it-abliterated-24 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49786566310722213}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | ymcki/gemma-2-2b-jpn-it-abliterated-17-18-24 | 7321bd04-6f20-427a-8219-0ff2e299cb01 | 0.0.1 | hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-17-18-24/1762652580.609858 | 1762652580.609859 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ymcki/gemma-2-2b-jpn-it-abliterated-17-18-24 | ymcki/gemma-2-2b-jpn-it-abliterated-17-18-24 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.505484337114412}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18 | 23800723-b5bd-4fc6-9d07-ca937c8680c6 | 0.0.1 | hfopenllm_v2/ymcki_gemma-2-2b-ORPO-jpn-it-abliterated-18/1762652580.6090298 | 1762652580.609031 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18 | ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4630945890237902}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | ymcki/gemma-2-2b-jpn-it-abliterated-18 | 78f235b0-fa98-48e2-bb03-9f7e9f986004 | 0.0.1 | hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-18/1762652580.610494 | 1762652580.610495 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ymcki/gemma-2-2b-jpn-it-abliterated-18 | ymcki/gemma-2-2b-jpn-it-abliterated-18 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5175246124726836}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO | 44b47789-f529-4bae-9e87-196abc325efc | 0.0.1 | hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-17-ORPO/1762652580.610075 | 1762652580.610076 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO | ymcki/gemma-2-2b-jpn-it-abliterated-17-ORPO | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47478468242042227}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | ymcki/gemma-2-2b-jpn-it-abliterated-18-ORPO | c91ab7d1-b36e-45ca-8f1e-ad9ef0c38100 | 0.0.1 | hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-18-ORPO/1762652580.610698 | 1762652580.610699 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ymcki/gemma-2-2b-jpn-it-abliterated-18-ORPO | ymcki/gemma-2-2b-jpn-it-abliterated-18-ORPO | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47423502972113984}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | ymcki/gemma-2-2b-jpn-it-abliterated-17 | 5958a61d-bf39-4de4-bfe1-6a6db2f37f55 | 0.0.1 | hfopenllm_v2/ymcki_gemma-2-2b-jpn-it-abliterated-17/1762652580.609628 | 1762652580.609628 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ymcki/gemma-2-2b-jpn-it-abliterated-17 | ymcki/gemma-2-2b-jpn-it-abliterated-17 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5081572449988254}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | Youlln/4PRYMMAL-GEMMA2-9B-SLERP | 06b75d54-4d17-4116-a4d5-0917eedb2dc4 | 0.0.1 | hfopenllm_v2/Youlln_4PRYMMAL-GEMMA2-9B-SLERP/1762652579.961175 | 1762652579.9611762 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | Youlln/4PRYMMAL-GEMMA2-9B-SLERP | Youlln/4PRYMMAL-GEMMA2-9B-SLERP | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2713766140507188}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | google/flan-t5-large | eb2e1202-9292-4f5e-a366-abc84897c66d | 0.0.1 | hfopenllm_v2/google_flan-t5-large/1762652580.173132 | 1762652580.1731331 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/flan-t5-large | google/flan-t5-large | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.22009490374428736}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "T5ForConditionalGeneration", "params_billions": 0.783} |
HF Open LLM v2 | google | jebish7/gemma-2-2b-it | 86206a02-3ab9-4a86-a00c-2900e8cd2e18 | 0.0.1 | hfopenllm_v2/jebish7_gemma-2-2b-it/1762652580.2824588 | 1762652580.2824588 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jebish7/gemma-2-2b-it | jebish7/gemma-2-2b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.12717035244263}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH",... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | jebish7/gemma-2-9b-it | 80a35d79-893b-439f-b100-a538a3c86974 | 0.0.1 | hfopenllm_v2/jebish7_gemma-2-9b-it/1762652580.282719 | 1762652580.28272 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jebish7/gemma-2-9b-it | jebish7/gemma-2-9b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1557467519514887}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | google/mt5-xxl | 38520cce-b3b6-4f22-a6a8-313f6181f5ea | 0.0.1 | hfopenllm_v2/google_mt5-xxl/1762652580.1791801 | 1762652580.1791801 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/mt5-xxl | google/mt5-xxl | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.23575668116154028}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "T5ForConditionalGeneration", "params_billions": 11.9} |
HF Open LLM v2 | google | princeton-nlp/gemma-2-9b-it-SimPO | 4285b38c-aba8-444b-9b0b-b265c7b1fef1 | 0.0.1 | hfopenllm_v2/princeton-nlp_gemma-2-9b-it-SimPO/1762652580.454763 | 1762652580.4547682 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | princeton-nlp/gemma-2-9b-it-SimPO | princeton-nlp/gemma-2-9b-it-SimPO | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3206857803960159}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | VAGOsolutions/SauerkrautLM-Gemma-2b | b002a274-9b4f-40ad-b0c7-e4efabbe431f | 0.0.1 | hfopenllm_v2/VAGOsolutions_SauerkrautLM-Gemma-2b/1762652579.941349 | 1762652579.94135 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | VAGOsolutions/SauerkrautLM-Gemma-2b | VAGOsolutions/SauerkrautLM-Gemma-2b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24752213017017072}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 2.506} |
HF Open LLM v2 | google | VAGOsolutions/SauerkrautLM-gemma-2-2b-it | b010858c-edb5-4e49-b5b6-72b06943ab2c | 0.0.1 | hfopenllm_v2/VAGOsolutions_SauerkrautLM-gemma-2-2b-it/1762652579.9427688 | 1762652579.94277 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | VAGOsolutions/SauerkrautLM-gemma-2-2b-it | VAGOsolutions/SauerkrautLM-gemma-2-2b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.13206625088099574}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | VAGOsolutions/SauerkrautLM-gemma-2-9b-it | 5395cbac-afe0-4936-b4eb-f554fcb5be75 | 0.0.1 | hfopenllm_v2/VAGOsolutions_SauerkrautLM-gemma-2-9b-it/1762652579.94298 | 1762652579.942981 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | VAGOsolutions/SauerkrautLM-gemma-2-9b-it | VAGOsolutions/SauerkrautLM-gemma-2-9b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3024009627787604}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | VAGOsolutions/SauerkrautLM-Gemma-7b | e66f4326-2585-4581-b45f-d9a81fb1576c | 0.0.1 | hfopenllm_v2/VAGOsolutions_SauerkrautLM-Gemma-7b/1762652579.9415941 | 1762652579.9415948 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | VAGOsolutions/SauerkrautLM-Gemma-7b | VAGOsolutions/SauerkrautLM-Gemma-7b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3406705319662939}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 8.538} |
HF Open LLM v2 | google | google/mt5-small | 0d958c7c-5cd9-459f-a0e9-235b5d41ae53 | 0.0.1 | hfopenllm_v2/google_mt5-small/1762652580.1787279 | 1762652580.178729 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/mt5-small | google/mt5-small | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.17180968718555653}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "MT5ForConditionalGeneration", "params_billions": 0.17} |
HF Open LLM v2 | google | ZHLiu627/zephyr-7b-gemma-rpo-avg | 6333359d-1cf7-4905-9a48-f8a8f7b46ed2 | 0.0.1 | hfopenllm_v2/ZHLiu627_zephyr-7b-gemma-rpo-avg/1762652579.9660559 | 1762652579.966057 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ZHLiu627/zephyr-7b-gemma-rpo-avg | ZHLiu627/zephyr-7b-gemma-rpo-avg | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30060350979844586}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "GemmaForCausalLM", "params_billions": 8.538} |
HF Open LLM v2 | google | anthracite-org/magnum-v3-9b-customgemma2 | 865b86aa-7b8d-4619-aa57-3c57cc4c7b51 | 0.0.1 | hfopenllm_v2/anthracite-org_magnum-v3-9b-customgemma2/1762652580.012768 | 1762652580.012769 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | anthracite-org/magnum-v3-9b-customgemma2 | anthracite-org/magnum-v3-9b-customgemma2 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1272955757390391}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | qq8933/OpenLongCoT-Base-Gemma2-2B | c945b9b5-7b46-4300-adcc-2d6c94df0ac1 | 0.0.1 | hfopenllm_v2/qq8933_OpenLongCoT-Base-Gemma2-2B/1762652580.488883 | 1762652580.488883 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | qq8933/OpenLongCoT-Base-Gemma2-2B | qq8933/OpenLongCoT-Base-Gemma2-2B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.1965141380426158}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 3.204} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v2-9B | ca1b9625-0112-4ebf-b1c3-d2dd217d50b2 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v2-9B/1762652580.315539 | 1762652580.31554 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v2-9B | lemon07r/Gemma-2-Ataraxy-v2-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.21362429464930827}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v4c-9B | 9499ec24-5be2-478c-b13e-3102d1555668 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v4c-9B/1762652580.318075 | 1762652580.318076 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v4c-9B | lemon07r/Gemma-2-Ataraxy-v4c-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6945282960323054}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v4a-Advanced-9B | b84aedba-7b87-445d-87c2-b029cb0038c3 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v4a-Advanced-9B/1762652580.317515 | 1762652580.317516 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v4a-Advanced-9B | lemon07r/Gemma-2-Ataraxy-v4a-Advanced-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7135123694020753}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v3i-9B | 53602c70-73d9-461b-b27a-24c6a1a538e5 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v3i-9B/1762652580.3165948 | 1762652580.316596 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v3i-9B | lemon07r/Gemma-2-Ataraxy-v3i-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4203047912871182}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v3j-9B | d435bd27-1c26-429d-8ac5-8fd8c591a9aa | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v3j-9B/1762652580.3168168 | 1762652580.316818 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v3j-9B | lemon07r/Gemma-2-Ataraxy-v3j-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4169326276501904}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v4-Advanced-9B | c0e95e3f-37a4-4b2f-a37b-37854546c241 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v4-Advanced-9B/1762652580.317157 | 1762652580.3171608 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v4-Advanced-9B | lemon07r/Gemma-2-Ataraxy-v4-Advanced-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7015474496558022}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-9B | 9ba72d50-4321-4383-8be9-286a56607624 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-9B/1762652580.31483 | 1762652580.314831 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-9B | lemon07r/Gemma-2-Ataraxy-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3008772279773224}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v2a-9B | 4fa1e172-f570-4a96-b53a-8ecf31854191 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v2a-9B/1762652580.315754 | 1762652580.315755 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v2a-9B | lemon07r/Gemma-2-Ataraxy-v2a-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.15946909755005606}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v2f-9B | fd59fb1c-3681-44d2-9172-b10891ae9c55 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v2f-9B/1762652580.315967 | 1762652580.315968 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v2f-9B | lemon07r/Gemma-2-Ataraxy-v2f-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.37911408396388246}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v3-Advanced-9B | 778a10b0-c537-4592-9dbb-2b0de07ced4c | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v3-Advanced-9B/1762652580.316169 | 1762652580.316169 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v3-Advanced-9B | lemon07r/Gemma-2-Ataraxy-v3-Advanced-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6601816513517467}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v4b-9B | 41f04f45-2f1d-42fd-87de-cc5e484cada2 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v4b-9B/1762652580.317803 | 1762652580.317804 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v4b-9B | lemon07r/Gemma-2-Ataraxy-v4b-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6878338364428604}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-Remix-9B | 29dfbb00-8760-46d8-bef8-d036870fb0c0 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-Remix-9B/1762652580.31531 | 1762652580.3153112 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-Remix-9B | lemon07r/Gemma-2-Ataraxy-Remix-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7083416446140685}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-Advanced-9B | 7806d1aa-b9e2-45bc-b89d-76e6c48dd3a0 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-Advanced-9B/1762652580.315091 | 1762652580.315092 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-Advanced-9B | lemon07r/Gemma-2-Ataraxy-Advanced-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5515964308036011}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v4d-9B | 7e6685d8-af21-4810-a9cc-edb296f4b937 | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v4d-9B/1762652580.318495 | 1762652580.318496 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v4d-9B | lemon07r/Gemma-2-Ataraxy-v4d-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7250029920610646}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | lemon07r/Gemma-2-Ataraxy-v3b-9B | d048e6ad-cc57-4ebe-8376-262564e86f0c | 0.0.1 | hfopenllm_v2/lemon07r_Gemma-2-Ataraxy-v3b-9B/1762652580.3163798 | 1762652580.316381 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | lemon07r/Gemma-2-Ataraxy-v3b-9B | lemon07r/Gemma-2-Ataraxy-v3b-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6809144181881852}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | ModelSpace/GemmaX2-28-9B-v0.1 | 6cb560eb-08f5-4430-8797-1116f1d2f56c | 0.0.1 | hfopenllm_v2/ModelSpace_GemmaX2-28-9B-v0.1/1762652579.76179 | 1762652579.761791 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | ModelSpace/GemmaX2-28-9B-v0.1 | ModelSpace/GemmaX2-28-9B-v0.1 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.003921816336210145}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on B... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | google/gemma-2-27b | 12f7d5a6-3f8b-49d8-9ca8-38774dbcca92 | 0.0.1 | hfopenllm_v2/google_gemma-2-27b/1762652580.175144 | 1762652580.175145 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2-27b | google/gemma-2-27b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.24752213017017072}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | google/gemma-2-2b | 53fb75b1-2d9f-4af3-a358-18bf5d4a9032 | 0.0.1 | hfopenllm_v2/google_gemma-2-2b/1762652580.1759539 | 1762652580.175955 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2-2b | google/gemma-2-2b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20176021844262113}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "InternLM2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | google/gemma-2-2b | 07e74f27-e0c3-448f-9a8c-a07ff8a73178 | 0.0.1 | hfopenllm_v2/google_gemma-2-2b/1762652580.175597 | 1762652580.1755981 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2-2b | google/gemma-2-2b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.19931226922343825}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "InternLM2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | google/gemma-2-2b-it | 64daa9ea-cf1e-4787-90cf-ed72c5e23afd | 0.0.1 | hfopenllm_v2/google_gemma-2-2b-it/1762652580.176172 | 1762652580.176194 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2-2b-it | google/gemma-2-2b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5668337788179807}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "InternLM2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | google/recurrentgemma-2b | 218a5d0f-5242-43c4-8166-81f5c09626bb | 0.0.1 | hfopenllm_v2/google_recurrentgemma-2b/1762652580.179393 | 1762652580.179394 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/recurrentgemma-2b | google/recurrentgemma-2b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3017028151970106}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "RecurrentGemmaForCausalLM", "params_billions": 2.683} |
HF Open LLM v2 | google | google/recurrentgemma-2b-it | a219b160-3dbd-4dcd-b39d-d12c6f9b1145 | 0.0.1 | hfopenllm_v2/google_recurrentgemma-2b-it/1762652580.17961 | 1762652580.179611 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/recurrentgemma-2b-it | google/recurrentgemma-2b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2949329999955673}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "RecurrentGemmaForCausalLM", "params_billions": 2.683} |
HF Open LLM v2 | google | google/codegemma-1.1-2b | 9d92e421-c458-4ad3-b9bf-45c0ca1b90cf | 0.0.1 | hfopenllm_v2/google_codegemma-1.1-2b/1762652580.172607 | 1762652580.172608 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/codegemma-1.1-2b | google/codegemma-1.1-2b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.22936253584932426}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 2.506} |
HF Open LLM v2 | google | google/gemma-1.1-7b-it | 6929c338-76a5-4386-9fa8-68e35a989a86 | 0.0.1 | hfopenllm_v2/google_gemma-1.1-7b-it/1762652580.1748302 | 1762652580.1748302 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-1.1-7b-it | google/gemma-1.1-7b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5039107346285633}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 8.538} |
HF Open LLM v2 | google | google/gemma-2-27b-it | 5bcf96ce-efd1-4f90-91a1-edd548de71ad | 0.0.1 | hfopenllm_v2/google_gemma-2-27b-it/1762652580.17537 | 1762652580.175371 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2-27b-it | google/gemma-2-27b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7977677008116243}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | google/gemma-2-9b-it | e8cef406-d6cc-48bd-872f-3d5b74bcf092 | 0.0.1 | hfopenllm_v2/google_gemma-2-9b-it/1762652580.177257 | 1762652580.177258 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2-9b-it | google/gemma-2-9b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7435626360279614}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.0} |
HF Open LLM v2 | google | google/gemma-2b-it | 50dffd1a-ddf5-40fd-a2c8-e5dd140af617 | 0.0.1 | hfopenllm_v2/google_gemma-2b-it/1762652580.17777 | 1762652580.17777 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2b-it | google/gemma-2b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.26902950837112194}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 2.506} |
HF Open LLM v2 | google | google/recurrentgemma-9b | 1ff3ab95-3007-4cbf-a146-5e8e4ae65404 | 0.0.1 | hfopenllm_v2/google_recurrentgemma-9b/1762652580.17984 | 1762652580.179841 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/recurrentgemma-9b | google/recurrentgemma-9b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31159434744256354}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "RecurrentGemmaForCausalLM", "params_billions": 9.0} |
HF Open LLM v2 | google | google/gemma-7b | 630e3cc0-fccc-41b3-b439-85a875dae401 | 0.0.1 | hfopenllm_v2/google_gemma-7b/1762652580.1780128 | 1762652580.178014 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-7b | google/gemma-7b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2659321710838353}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 8.538} |
HF Open LLM v2 | google | google/gemma-2-9b | 2ac50111-a850-4bd2-8136-c373990742a5 | 0.0.1 | hfopenllm_v2/google_gemma-2-9b/1762652580.177011 | 1762652580.177012 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2-9b | google/gemma-2-9b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20398320899657355}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.0} |
HF Open LLM v2 | google | google/gemma-2b | 2dd86ebc-0253-4801-ac99-2bb3494ad29b | 0.0.1 | hfopenllm_v2/google_gemma-2b/1762652580.177512 | 1762652580.177513 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2b | google/gemma-2b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20375825033134307}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 2.506} |
HF Open LLM v2 | google | google/gemma-7b-it | 30146048-ee0f-431d-b3e7-8c066c820740 | 0.0.1 | hfopenllm_v2/google_gemma-7b-it/1762652580.178242 | 1762652580.1782432 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-7b-it | google/gemma-7b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3868324933398937}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 8.538} |
HF Open LLM v2 | google | google/gemma-1.1-2b-it | 5ed676b6-4aff-4d71-a91a-6d5d9feeb28f | 0.0.1 | hfopenllm_v2/google_gemma-1.1-2b-it/1762652580.1745641 | 1762652580.174565 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-1.1-2b-it | google/gemma-1.1-2b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.30674831668860847}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "GemmaForCausalLM", "params_billions": 2.506} |
HF Open LLM v2 | google | google/recurrentgemma-9b-it | c7095b76-2d50-467b-a8d9-d7a277f1f14c | 0.0.1 | hfopenllm_v2/google_recurrentgemma-9b-it/1762652580.180049 | 1762652580.18005 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/recurrentgemma-9b-it | google/recurrentgemma-9b-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5010383560065071}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "RecurrentGemmaForCausalLM", "params_billions": 9.0} |
HF Open LLM v2 | google | google/gemma-2-2b-jpn-it | 251b93fa-6f12-41bc-85c8-ded52e1a0d2d | 0.0.1 | hfopenllm_v2/google_gemma-2-2b-jpn-it/1762652580.1767948 | 1762652580.176796 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2-2b-jpn-it | google/gemma-2-2b-jpn-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5288401441508531}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | google/gemma-2-2b-jpn-it | a09fdbce-489c-4d14-a05f-7663121bece7 | 0.0.1 | hfopenllm_v2/google_gemma-2-2b-jpn-it/1762652580.176506 | 1762652580.176507 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/gemma-2-2b-jpn-it | google/gemma-2-2b-jpn-it | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5077826832803628}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | google/flan-ul2 | 99941572-3e23-467c-97df-dfe1a2aa9805 | 0.0.1 | hfopenllm_v2/google_flan-ul2/1762652580.174251 | 1762652580.174251 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/flan-ul2 | google/flan-ul2 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.23925406809487715}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "T5ForConditionalGeneration", "params_billions": 19.46} |
HF Open LLM v2 | google | wzhouad/gemma-2-9b-it-WPO-HB | 70fe199f-6c81-4d99-a595-208b7abc321f | 0.0.1 | hfopenllm_v2/wzhouad_gemma-2-9b-it-WPO-HB/1762652580.596365 | 1762652580.5963662 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | wzhouad/gemma-2-9b-it-WPO-HB | wzhouad/gemma-2-9b-it-WPO-HB | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5437029304467702}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | UCLA-AGI/Gemma-2-9B-It-SPPO-Iter2 | fa584f01-69eb-4ecc-9f0d-049b6bfb05c8 | 0.0.1 | hfopenllm_v2/UCLA-AGI_Gemma-2-9B-It-SPPO-Iter2/1762652579.936279 | 1762652579.93628 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | UCLA-AGI/Gemma-2-9B-It-SPPO-Iter2 | UCLA-AGI/Gemma-2-9B-It-SPPO-Iter2 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3100196367859502}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3 | f318d457-d295-4447-9222-0b0d92708b5d | 0.0.1 | hfopenllm_v2/UCLA-AGI_Gemma-2-9B-It-SPPO-Iter3/1762652579.9364889 | 1762652579.93649 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3 | UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31671409637539505}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | UCLA-AGI/Gemma-2-9B-It-SPPO-Iter1 | 687769ed-44e9-4f3d-aee6-2dc4e98dd7ee | 0.0.1 | hfopenllm_v2/UCLA-AGI_Gemma-2-9B-It-SPPO-Iter1/1762652579.936019 | 1762652579.93602 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | UCLA-AGI/Gemma-2-9B-It-SPPO-Iter1 | UCLA-AGI/Gemma-2-9B-It-SPPO-Iter1 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.308221075634871}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | hotmailuser/Gemma2Crono-27B | 501e2a2c-e32c-455e-8e5f-f8bde053fddc | 0.0.1 | hfopenllm_v2/hotmailuser_Gemma2Crono-27B/1762652580.193866 | 1762652580.193866 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/Gemma2Crono-27B | hotmailuser/Gemma2Crono-27B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7086164709637096}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | hotmailuser/Gemma2atlas-27B | c9020f27-9175-4f12-a108-6cbff1c0cb22 | 0.0.1 | hfopenllm_v2/hotmailuser_Gemma2atlas-27B/1762652580.1943119 | 1762652580.194313 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/Gemma2atlas-27B | hotmailuser/Gemma2atlas-27B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7213560020744957}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | hotmailuser/Gemma2magnum-27b | 0ad192a1-b33f-4362-a21d-ccc590986c5c | 0.0.1 | hfopenllm_v2/hotmailuser_Gemma2magnum-27b/1762652580.1945128 | 1762652580.194514 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/Gemma2magnum-27b | hotmailuser/Gemma2magnum-27b | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5050599077115387}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | hotmailuser/Gemma2SimPO-27B | 433a8abf-8ff7-40bb-a4d0-654efdb6bf86 | 0.0.1 | hfopenllm_v2/hotmailuser_Gemma2SimPO-27B/1762652580.194106 | 1762652580.1941068 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | hotmailuser/Gemma2SimPO-27B | hotmailuser/Gemma2SimPO-27B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7222303488078299}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | sequelbox/gemma-2-9B-MOTH | 4bdefb85-2413-43b7-8938-869ad0cff58f | 0.0.1 | hfopenllm_v2/sequelbox_gemma-2-9B-MOTH/1762652580.5126731 | 1762652580.512674 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | sequelbox/gemma-2-9B-MOTH | sequelbox/gemma-2-9B-MOTH | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.20588150551647405}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | recoilme/recoilme-gemma-2-9B-v0.5 | 8fe5a1e8-1491-4e64-8aed-32e73f2dae6e | 0.0.1 | hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.5/1762652580.4931269 | 1762652580.493134 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | recoilme/recoilme-gemma-2-9B-v0.5 | recoilme/recoilme-gemma-2-9B-v0.5 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7664186580495308}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | recoilme/recoilme-gemma-2-9B-v0.3 | 8d3bd687-89f5-4d62-af46-93646aea4341 | 0.0.1 | hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.3/1762652580.492666 | 1762652580.492667 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | recoilme/recoilme-gemma-2-9B-v0.3 | recoilme/recoilme-gemma-2-9B-v0.3 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.57607592299543}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH",... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | recoilme/recoilme-gemma-2-9B-v0.3 | 47cfe707-ba31-4c9b-aa15-9ab8b566e206 | 0.0.1 | hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.3/1762652580.492416 | 1762652580.492416 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | recoilme/recoilme-gemma-2-9B-v0.3 | recoilme/recoilme-gemma-2-9B-v0.3 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.743937197746424}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | recoilme/recoilme-gemma-2-9B-v0.2 | 6a15378c-36cc-4f5e-b184-5a19a6fbb192 | 0.0.1 | hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.2/1762652580.492019 | 1762652580.49202 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | recoilme/recoilme-gemma-2-9B-v0.2 | recoilme/recoilme-gemma-2-9B-v0.2 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7591745457608035}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | recoilme/recoilme-gemma-2-9B-v0.2 | 5826c93f-3642-44cf-b385-4a5ab5103086 | 0.0.1 | hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.2/1762652580.4922318 | 1762652580.492233 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | recoilme/recoilme-gemma-2-9B-v0.2 | recoilme/recoilme-gemma-2-9B-v0.2 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2746989100032359}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | recoilme/recoilme-gemma-2-9B-v0.4 | 28eef1b7-a83e-49c9-8f11-ef9e4ae7e1ce | 0.0.1 | hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.4/1762652580.4928808 | 1762652580.492882 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | recoilme/recoilme-gemma-2-9B-v0.4 | recoilme/recoilme-gemma-2-9B-v0.4 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2561891337207498}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp | 0a685d8f-38c7-4521-9613-7b36ad1cac73 | 0.0.1 | hfopenllm_v2/recoilme_Gemma-2-Ataraxy-Gemmasutra-9B-slerp/1762652580.491603 | 1762652580.491603 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp | recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.28536505361330156}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp | 054a662a-e425-448c-9556-6998833e51ff | 0.0.1 | hfopenllm_v2/recoilme_Gemma-2-Ataraxy-Gemmasutra-9B-slerp/1762652580.491333 | 1762652580.491333 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp | recoilme/Gemma-2-Ataraxy-Gemmasutra-9B-slerp | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7648949232480928}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | recoilme/recoilme-gemma-2-9B-v0.1 | d31a41b0-6500-4e1b-8435-b9d3e9725c02 | 0.0.1 | hfopenllm_v2/recoilme_recoilme-gemma-2-9B-v0.1/1762652580.491797 | 1762652580.491798 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | recoilme/recoilme-gemma-2-9B-v0.1 | recoilme/recoilme-gemma-2-9B-v0.1 | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.751506004069203}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "float16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | allknowingroger/Gemma2Slerp2-27B | 1f2c33e8-2d7b-4bd5-81e8-1c9bcae0ae8f | 0.0.1 | hfopenllm_v2/allknowingroger_Gemma2Slerp2-27B/1762652579.986531 | 1762652579.9865322 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/Gemma2Slerp2-27B | allknowingroger/Gemma2Slerp2-27B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7545534736720789}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | allknowingroger/Gemma2Slerp2-2.6B | eeb46285-0c8d-43b7-9b6d-e86c24064fde | 0.0.1 | hfopenllm_v2/allknowingroger_Gemma2Slerp2-2.6B/1762652579.98633 | 1762652579.98633 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/Gemma2Slerp2-2.6B | allknowingroger/Gemma2Slerp2-2.6B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5747272791748117}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | allknowingroger/Gemma2Slerp1-27B | 42d79295-bdb0-411d-b1b0-5cff954e925c | 0.0.1 | hfopenllm_v2/allknowingroger_Gemma2Slerp1-27B/1762652579.986121 | 1762652579.986122 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/Gemma2Slerp1-27B | allknowingroger/Gemma2Slerp1-27B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7186332265056716}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | allknowingroger/GemmaSlerp4-10B | 32e38c82-d412-4888-9d9d-f89aef0989fd | 0.0.1 | hfopenllm_v2/allknowingroger_GemmaSlerp4-10B/1762652579.9875991 | 1762652579.9875998 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/GemmaSlerp4-10B | allknowingroger/GemmaSlerp4-10B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7326216660682544}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | allknowingroger/GemmaSlerp-9B | 3aed9fd2-45bd-4568-8885-7fc2370bb26d | 0.0.1 | hfopenllm_v2/allknowingroger_GemmaSlerp-9B/1762652579.987181 | 1762652579.9871821 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/GemmaSlerp-9B | allknowingroger/GemmaSlerp-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.704320092909037}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | allknowingroger/GemmaSlerp5-10B | e325b56f-4306-4e37-adc5-c09b300a8c30 | 0.0.1 | hfopenllm_v2/allknowingroger_GemmaSlerp5-10B/1762652579.9878101 | 1762652579.987811 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/GemmaSlerp5-10B | allknowingroger/GemmaSlerp5-10B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7353444416370785}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 10.159} |
HF Open LLM v2 | google | allknowingroger/GemmaSlerp2-9B | 99333370-c7d5-4763-b3a4-14adde0fab9e | 0.0.1 | hfopenllm_v2/allknowingroger_GemmaSlerp2-9B/1762652579.987394 | 1762652579.987395 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/GemmaSlerp2-9B | allknowingroger/GemmaSlerp2-9B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7281003293483512}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 9.242} |
HF Open LLM v2 | google | allknowingroger/Gemma2Slerp3-27B | 648810d4-4dd5-48c7-a4d7-b3d9d2f3f3f2 | 0.0.1 | hfopenllm_v2/allknowingroger_Gemma2Slerp3-27B/1762652579.986752 | 1762652579.986753 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/Gemma2Slerp3-27B | allknowingroger/Gemma2Slerp3-27B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7426384216102164}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | allknowingroger/Gemma2Slerp1-2.6B | e52ac657-26a3-499a-949f-bf2a0b620d8e | 0.0.1 | hfopenllm_v2/allknowingroger_Gemma2Slerp1-2.6B/1762652579.985875 | 1762652579.985876 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/Gemma2Slerp1-2.6B | allknowingroger/Gemma2Slerp1-2.6B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5354348683714766}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 2.614} |
HF Open LLM v2 | google | allknowingroger/Gemma2Slerp4-27B | f94f3bf1-cf85-4673-a5cf-368f250233e4 | 0.0.1 | hfopenllm_v2/allknowingroger_Gemma2Slerp4-27B/1762652579.986965 | 1762652579.9869661 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/Gemma2Slerp4-27B | allknowingroger/Gemma2Slerp4-27B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7496575752337131}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | allknowingroger/GemmaStock1-27B | 0b19d8bb-1952-4515-8d29-e55e1106e92b | 0.0.1 | hfopenllm_v2/allknowingroger_GemmaStock1-27B/1762652579.9880252 | 1762652579.9880252 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | allknowingroger/GemmaStock1-27B | allknowingroger/GemmaStock1-27B | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7509064836855099}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "Gemma2ForCausalLM", "params_billions": 27.227} |
HF Open LLM v2 | google | google/mt5-xl | 5abb3ce9-6ad4-4dfa-8bca-81ec6cb84426 | 0.0.1 | hfopenllm_v2/google_mt5-xl/1762652580.17897 | 1762652580.1789708 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | google/mt5-xl | google/mt5-xl | google | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.19596448534333347}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "MT5ForConditionalGeneration", "params_billions": 3.23} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.