_leaderboard stringclasses 1
value | _developer stringclasses 559
values | _model stringlengths 9 102 | _uuid stringlengths 36 36 | schema_version stringclasses 1
value | evaluation_id stringlengths 35 133 | retrieved_timestamp stringlengths 13 18 | source_data stringclasses 1
value | evaluation_source_name stringclasses 1
value | evaluation_source_type stringclasses 1
value | source_organization_name stringclasses 1
value | source_organization_url null | source_organization_logo_url null | evaluator_relationship stringclasses 1
value | model_name stringlengths 4 102 | model_id stringlengths 9 102 | model_developer stringclasses 559
values | model_inference_platform stringclasses 1
value | evaluation_results stringlengths 1.35k 1.41k | additional_details stringclasses 660
values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
HF Open LLM v2 | icefog72 | icefog72/Ice0.62-18.01-RP | 0c5bb530-f59b-4097-8a79-9e4f524385a2 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.62-18.01-RP/1762652580.21767 | 1762652580.217671 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.62-18.01-RP | icefog72/Ice0.62-18.01-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.536733644507684}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/IceSakeV8RP-7b | dbeb9a8a-53c5-472b-a4b1-1aa0582f8486 | 0.0.1 | hfopenllm_v2/icefog72_IceSakeV8RP-7b/1762652580.2249868 | 1762652580.224988 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/IceSakeV8RP-7b | icefog72/IceSakeV8RP-7b | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6085741388404988}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.52.1-16.01-RP | 6bfbd9d6-b376-4169-8e6a-2c3210040e97 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.52.1-16.01-RP/1762652580.21567 | 1762652580.215671 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.52.1-16.01-RP | icefog72/Ice0.52.1-16.01-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45492626231731803}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.27-06.11-RP | 314c9c7e-0c13-4f6b-be25-d2a2cbc25e9b | 0.0.1 | hfopenllm_v2/icefog72_Ice0.27-06.11-RP/1762652580.211702 | 1762652580.211702 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.27-06.11-RP | icefog72/Ice0.27-06.11-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49182059158588104}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.51-16.01-RP | 7a137ac4-8445-4c1a-9203-abc5f4131213 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.51-16.01-RP/1762652580.214901 | 1762652580.214902 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.51-16.01-RP | icefog72/Ice0.51-16.01-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4430610779398662}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/IceDrunkCherryRP-7b | c0e3f4ee-52dc-45c3-844a-8cc4e4520f24 | 0.0.1 | hfopenllm_v2/icefog72_IceDrunkCherryRP-7b/1762652580.222923 | 1762652580.222924 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/IceDrunkCherryRP-7b | icefog72/IceDrunkCherryRP-7b | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48982255969715904}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.16-02.10-RP | 824cb85d-e7a0-421a-994b-c0b178ab8e56 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.16-02.10-RP/1762652580.211284 | 1762652580.211284 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.16-02.10-RP | icefog72/Ice0.16-02.10-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5069083365470286}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.78-02.02-RP | ec943fa1-b138-46e8-b1ae-c9a476c73ed1 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.78-02.02-RP/1762652580.221266 | 1762652580.221267 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.78-02.02-RP | icefog72/Ice0.78-02.02-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.405292401937969}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/IceTea21EnergyDrinkRPV13-DPOv3.5 | f4d3a112-d529-48f8-a99e-85e9eb02e0c1 | 0.0.1 | hfopenllm_v2/icefog72_IceTea21EnergyDrinkRPV13-DPOv3.5/1762652580.2254012 | 1762652580.225402 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/IceTea21EnergyDrinkRPV13-DPOv3.5 | icefog72/IceTea21EnergyDrinkRPV13-DPOv3.5 | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48709978412833504}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/IceSakeV4RP-7b | 93b5850f-74d0-45cd-977e-5bf6e4dc5d8d | 0.0.1 | hfopenllm_v2/icefog72_IceSakeV4RP-7b/1762652580.224551 | 1762652580.224552 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/IceSakeV4RP-7b | icefog72/IceSakeV4RP-7b | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4634192830578421}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.51.1-16.01-RP | 859a9706-f73b-4426-9c5a-052625d62f5b | 0.0.1 | hfopenllm_v2/icefog72_Ice0.51.1-16.01-RP/1762652580.215148 | 1762652580.2151492 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.51.1-16.01-RP | icefog72/Ice0.51.1-16.01-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4573243438520902}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.70-25.01-RP | e109acd0-c7e3-4a9f-8e06-c428b95acc83 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.70-25.01-RP/1762652580.2196732 | 1762652580.219674 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.70-25.01-RP | icefog72/Ice0.70-25.01-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.549797869652522}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.73-01.02-RP | ba7bf09f-b7a1-4fd4-b262-4929a81da34a | 0.0.1 | hfopenllm_v2/icefog72_Ice0.73-01.02-RP/1762652580.220075 | 1762652580.220076 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.73-01.02-RP | icefog72/Ice0.73-01.02-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.529164838184905}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.31-08.11-RP | 1fc072c6-ad31-4151-8420-7402b565510d | 0.0.1 | hfopenllm_v2/icefog72_Ice0.31-08.11-RP/1762652580.212094 | 1762652580.212095 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.31-08.11-RP | icefog72/Ice0.31-08.11-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5145768782386291}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.39-19.11-RP | 780c711f-774b-499e-881e-25dba76273a1 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.39-19.11-RP/1762652580.2133162 | 1762652580.2133162 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.39-19.11-RP | icefog72/Ice0.39-19.11-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47565902915375646}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.17-03.10-RP | 2faf039c-9c8e-46db-8472-6b741c451bf1 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.17-03.10-RP/1762652580.211494 | 1762652580.211495 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.17-03.10-RP | icefog72/Ice0.17-03.10-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5123538876846767}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.65-25.01-RP | fa5d2148-c45b-4266-a6a0-11b471273f75 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.65-25.01-RP/1762652580.2184708 | 1762652580.218472 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.65-25.01-RP | icefog72/Ice0.65-25.01-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5029366525264077}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.77-02.02-RP | 0eebefc6-138f-4af5-a8b6-a35c798a38cb | 0.0.1 | hfopenllm_v2/icefog72_Ice0.77-02.02-RP/1762652580.221007 | 1762652580.2210078 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.77-02.02-RP | icefog72/Ice0.77-02.02-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5309633993359841}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/IceTea21EnergyDrinkRPV13-DPOv3 | 4b4a9630-c942-445e-b396-4a988d489aa7 | 0.0.1 | hfopenllm_v2/icefog72_IceTea21EnergyDrinkRPV13-DPOv3/1762652580.225198 | 1762652580.2251992 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/IceTea21EnergyDrinkRPV13-DPOv3 | icefog72/IceTea21EnergyDrinkRPV13-DPOv3 | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5263423272472595}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.37-18.11-RP | 774c0461-5e81-436a-9347-7a4cc15ca019 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.37-18.11-RP/1762652580.212915 | 1762652580.212916 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.37-18.11-RP | icefog72/Ice0.37-18.11-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4972162750391184}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.70.1-01.02-RP | ee088f70-5734-4951-8bc0-e0579a053fd2 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.70.1-01.02-RP/1762652580.219877 | 1762652580.219877 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.70.1-01.02-RP | icefog72/Ice0.70.1-01.02-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5069582042314393}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/IceDrinkByFrankensteinV3RP | 39325b65-ad12-44ef-a1bf-ffe9e870ced8 | 0.0.1 | hfopenllm_v2/icefog72_IceDrinkByFrankensteinV3RP/1762652580.222236 | 1762652580.222236 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/IceDrinkByFrankensteinV3RP | icefog72/IceDrinkByFrankensteinV3RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4974911013887596}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.62.1-24.01-RP | 26ba869e-ae3b-44ef-a215-f94e4e4cb1fc | 0.0.1 | hfopenllm_v2/icefog72_Ice0.62.1-24.01-RP/1762652580.2178729 | 1762652580.2178729 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.62.1-24.01-RP | icefog72/Ice0.62.1-24.01-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5181740005407873}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.40-20.11-RP | 5220bee5-74d3-4730-9fee-4ca488e1a37e | 0.0.1 | hfopenllm_v2/icefog72_Ice0.40-20.11-RP/1762652580.2136111 | 1762652580.213614 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.40-20.11-RP | icefog72/Ice0.40-20.11-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4762585495374495}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/IceDrinkNameGoesHereRP-7b-Model_Stock | b0aaf6e9-ffe3-4de9-b3f5-c33d52b59ed2 | 0.0.1 | hfopenllm_v2/icefog72_IceDrinkNameGoesHereRP-7b-Model_Stock/1762652580.2224698 | 1762652580.2224698 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/IceDrinkNameGoesHereRP-7b-Model_Stock | icefog72/IceDrinkNameGoesHereRP-7b-Model_Stock | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49684171332065585}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.76-02.02-RP | 701743bb-1ddf-4810-824a-38959d4a0e02 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.76-02.02-RP/1762652580.220735 | 1762652580.220737 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.76-02.02-RP | icefog72/Ice0.76-02.02-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45290274250100837}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.60.1-18.01-RP | 8a14ed64-1408-469e-ab8d-05c897904d20 | 0.0.1 | hfopenllm_v2/icefog72_Ice0.60.1-18.01-RP/1762652580.217258 | 1762652580.217259 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.60.1-18.01-RP | icefog72/Ice0.60.1-18.01-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5187735209244804}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.80-03.02-RP | 847b4e14-a07c-45ed-b2eb-ecea0f80147b | 0.0.1 | hfopenllm_v2/icefog72_Ice0.80-03.02-RP/1762652580.2214909 | 1762652580.221492 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.80-03.02-RP | icefog72/Ice0.80-03.02-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5516462984880118}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.54-17.01-RP | 94d01e56-d7d5-4680-b577-ebcc0198ca0c | 0.0.1 | hfopenllm_v2/icefog72_Ice0.54-17.01-RP/1762652580.2162719 | 1762652580.2162728 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.54-17.01-RP | icefog72/Ice0.54-17.01-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4378903531518593}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/IceNalyvkaRP-7b | 95dd235d-6930-48fd-8594-5acb0110be29 | 0.0.1 | hfopenllm_v2/icefog72_IceNalyvkaRP-7b/1762652580.224114 | 1762652580.224115 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/IceNalyvkaRP-7b | icefog72/IceNalyvkaRP-7b | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.549797869652522}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | icefog72 | icefog72/Ice0.34b-14.11-RP | ed2a47c3-06c7-451b-94cd-8cd42be2ca9c | 0.0.1 | hfopenllm_v2/icefog72_Ice0.34b-14.11-RP/1762652580.2124958 | 1762652580.212497 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | icefog72/Ice0.34b-14.11-RP | icefog72/Ice0.34b-14.11-RP | icefog72 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47620868185303883}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242} |
HF Open LLM v2 | bluuwhale | bluuwhale/L3-SthenoMaid-8B-V1 | 44dd13bc-56f0-4dd1-90d0-bb411239109a | 0.0.1 | hfopenllm_v2/bluuwhale_L3-SthenoMaid-8B-V1/1762652580.035146 | 1762652580.035147 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | bluuwhale/L3-SthenoMaid-8B-V1 | bluuwhale/L3-SthenoMaid-8B-V1 | bluuwhale | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7344700949037443}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-Orochi-v2-Experiment | db8eedcc-1dcf-47af-9c2b-a72da97146ca | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-Orochi-v2-Experiment/1762652579.5591779 | 1762652579.559179 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-Orochi-v2-Experiment | DoppelReflEx/MN-12B-Mimicore-Orochi-v2-Experiment | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2842413684579139}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-Nocturne | 6a21892f-1d11-4c59-8894-8800822b2e72 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-Nocturne/1762652579.558723 | 1762652579.5587242 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-Nocturne | DoppelReflEx/MN-12B-Mimicore-Nocturne | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3956502081144696}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-4 | d7303703-f33e-430b-813d-998c95dbdb67 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-WhiteSnake-v2-Experiment-4/1762652579.560668 | 1762652579.560668 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-4 | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-4 | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42405151664250856}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-LilithFrame-Experiment-2 | 630c100f-c88d-42a7-9614-bd9a958eab2b | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-LilithFrame-Experiment-2/1762652579.5578592 | 1762652579.5578601 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-LilithFrame-Experiment-2 | DoppelReflEx/MN-12B-LilithFrame-Experiment-2 | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4299469851106176}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/L3-8B-WolfCore | c6771d5c-acaf-4b17-96b4-abf3b75bc68f | 0.0.1 | hfopenllm_v2/DoppelReflEx_L3-8B-WolfCore/1762652579.556399 | 1762652579.5564 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/L3-8B-WolfCore | DoppelReflEx/L3-8B-WolfCore | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4021950646506824}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MiniusLight-24B-v1b-test | 2917ef74-c8cb-4255-8bda-76280fbe7c64 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MiniusLight-24B-v1b-test/1762652579.561931 | 1762652579.561932 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MiniusLight-24B-v1b-test | DoppelReflEx/MiniusLight-24B-v1b-test | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.37911408396388246}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 23.572} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-1 | aa2478d9-59bd-458b-abee-5669aa6280df | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-WhiteSnake-v2-Experiment-1/1762652579.5600362 | 1762652579.5600362 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-1 | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-1 | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.39090391272933595}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Unleashed-Twilight | 00f0fe96-4a06-46e7-88d8-368b86bcdb06 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Unleashed-Twilight/1762652579.560919 | 1762652579.56092 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Unleashed-Twilight | DoppelReflEx/MN-12B-Unleashed-Twilight | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3505121965274361}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-WhiteSnake | 8aa34df4-8347-4f2d-98a0-7ec58bd62e43 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-WhiteSnake/1762652579.55982 | 1762652579.5598211 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-WhiteSnake | DoppelReflEx/MN-12B-Mimicore-WhiteSnake | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44376033369238066}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-2 | 66bd7a21-6f85-49b5-bc01-3f52ed8d1c64 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-WhiteSnake-v2-Experiment-2/1762652579.560246 | 1762652579.560246 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-2 | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-2 | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31239333856389934}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-Orochi | f1bfef73-3586-4f9d-80ca-71b0fb00aadd | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-Orochi/1762652579.558937 | 1762652579.558938 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-Orochi | DoppelReflEx/MN-12B-Mimicore-Orochi | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4620451513096362}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/L3-8B-R1-WolfCore-V1.5-test | 3c4058cd-238b-4b01-870d-8693f5ce1b8f | 0.0.1 | hfopenllm_v2/DoppelReflEx_L3-8B-R1-WolfCore-V1.5-test/1762652579.556192 | 1762652579.556193 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/L3-8B-R1-WolfCore-V1.5-test | DoppelReflEx/L3-8B-R1-WolfCore-V1.5-test | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3955006050612375}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/L3-8B-R1-WolfCore | 6d8d63c0-ad69-4224-8250-b1664f6abbcf | 0.0.1 | hfopenllm_v2/DoppelReflEx_L3-8B-R1-WolfCore/1762652579.555949 | 1762652579.5559502 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/L3-8B-R1-WolfCore | DoppelReflEx/L3-8B-R1-WolfCore | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3775404814780339}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Kakigori | 2f19082b-8377-4f63-8c5f-1aa25071a240 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Kakigori/1762652579.5572648 | 1762652579.557266 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Kakigori | DoppelReflEx/MN-12B-Kakigori | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.359329911302012}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-Orochi-v4-Experiment | e4e71999-6f83-4745-8a9d-66e711e39ac3 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-Orochi-v4-Experiment/1762652579.559606 | 1762652579.559606 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-Orochi-v4-Experiment | DoppelReflEx/MN-12B-Mimicore-Orochi-v4-Experiment | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4320702402957486}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-LilithFrame-Experiment-3 | 37292ca7-9e82-4c80-bc6e-bc7e1be7a95e | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-LilithFrame-Experiment-3/1762652579.558079 | 1762652579.558079 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-LilithFrame-Experiment-3 | DoppelReflEx/MN-12B-LilithFrame-Experiment-3 | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4127858526487498}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-Orochi-v3-Experiment | 8198ab16-4a8b-4da9-8e8a-d1e3beb02839 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-Orochi-v3-Experiment/1762652579.559391 | 1762652579.559392 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-Orochi-v3-Experiment | DoppelReflEx/MN-12B-Mimicore-Orochi-v3-Experiment | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4101628124487471}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-GreenSnake | 9b9eb072-4120-4a6a-a565-27136e617f10 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-GreenSnake/1762652579.5585039 | 1762652579.558505 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-GreenSnake | DoppelReflEx/MN-12B-Mimicore-GreenSnake | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47800724300411795}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MiniusLight-24B | 2ec36e2e-0fba-4c6a-b9d0-fe57e7d708ef | 0.0.1 | hfopenllm_v2/DoppelReflEx_MiniusLight-24B/1762652579.561418 | 1762652579.561419 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MiniusLight-24B | DoppelReflEx/MiniusLight-24B | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.25766410900854175}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 23.572} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-3 | 1a3eefa6-7b3d-4541-93b0-8fe86f6bf038 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-WhiteSnake-v2-Experiment-3/1762652579.56046 | 1762652579.560461 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-3 | DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-3 | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4302218114602588}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MiniusLight-24B-test | e6031abf-1ae2-431c-8247-3124fff41d17 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MiniusLight-24B-test/1762652579.5616372 | 1762652579.5616379 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MiniusLight-24B-test | DoppelReflEx/MiniusLight-24B-test | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.03936776641533354}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 23.572} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-FoxFrame-test | ef5bb4eb-0875-4cc5-8e27-b59ffbd2e477 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-FoxFrame-test/1762652579.556618 | 1762652579.556619 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-FoxFrame-test | DoppelReflEx/MN-12B-FoxFrame-test | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42220308780701876}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-FoxFrame3-test | 35351894-ea9d-456b-ab9a-c98686948e6b | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-FoxFrame3-test/1762652579.557049 | 1762652579.5570502 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-FoxFrame3-test | DoppelReflEx/MN-12B-FoxFrame3-test | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43231957871780213}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MiniusLight-24B-v1d-test | af67712e-7436-4703-ac22-9878dd8e190a | 0.0.1 | hfopenllm_v2/DoppelReflEx_MiniusLight-24B-v1d-test/1762652579.5624058 | 1762652579.5624058 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MiniusLight-24B-v1d-test | DoppelReflEx/MiniusLight-24B-v1d-test | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.40324339419407174}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 23.572} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-FoxFrame2-test | e46698de-8b2d-4b3c-b482-8cc8a3665eac | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-FoxFrame2-test/1762652579.556837 | 1762652579.5568378 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-FoxFrame2-test | DoppelReflEx/MN-12B-FoxFrame2-test | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43189514931492884}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-LilithFrame | a04a8775-8b4d-4608-9692-47af9f7ed5a7 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-LilithFrame/1762652579.557468 | 1762652579.557469 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-LilithFrame | DoppelReflEx/MN-12B-LilithFrame | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4509545782966972}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-LilithFrame | 3d68e2fb-06cc-43b9-830b-f1cd02f12166 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-LilithFrame/1762652579.557674 | 1762652579.5576751 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-LilithFrame | DoppelReflEx/MN-12B-LilithFrame | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43604192431636946}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-LilithFrame-Experiment-4 | ecc18f9c-c495-4ae6-8fd8-b2f84fb453ac | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-LilithFrame-Experiment-4/1762652579.5582879 | 1762652579.5582888 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-LilithFrame-Experiment-4 | DoppelReflEx/MN-12B-LilithFrame-Experiment-4 | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3981480250180632}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MiniusLight-24B-v1c-test | 23a21492-0897-44b4-a046-cf93fa8c2a64 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MiniusLight-24B-v1c-test/1762652579.562173 | 1762652579.5621738 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MiniusLight-24B-v1c-test | DoppelReflEx/MiniusLight-24B-v1c-test | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.37858881102142317}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 23.572} |
HF Open LLM v2 | DoppelReflEx | DoppelReflEx/MN-12B-WolFrame | 3bb96e7a-6c09-4b9e-8f2b-0b525c2ebeb3 | 0.0.1 | hfopenllm_v2/DoppelReflEx_MN-12B-WolFrame/1762652579.5611808 | 1762652579.561182 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | DoppelReflEx/MN-12B-WolFrame | DoppelReflEx/MN-12B-WolFrame | DoppelReflEx | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4397387819873491}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248} |
HF Open LLM v2 | sethuiyer | sethuiyer/Llamazing-3.1-8B-Instruct | 9065a7df-dab7-4e3b-bbc5-01f2908c37b3 | 0.0.1 | hfopenllm_v2/sethuiyer_Llamazing-3.1-8B-Instruct/1762652580.513854 | 1762652580.513855 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | sethuiyer/Llamazing-3.1-8B-Instruct | sethuiyer/Llamazing-3.1-8B-Instruct | sethuiyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5711301568726534}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | sethuiyer | sethuiyer/Llama-3.1-8B-Experimental-1206-Instruct | 49334550-08eb-49a2-9cea-f90f22533ab1 | 0.0.1 | hfopenllm_v2/sethuiyer_Llama-3.1-8B-Experimental-1206-Instruct/1762652580.512954 | 1762652580.512954 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | sethuiyer/Llama-3.1-8B-Experimental-1206-Instruct | sethuiyer/Llama-3.1-8B-Experimental-1206-Instruct | sethuiyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6967014189018471}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | sethuiyer | sethuiyer/Llama-3.1-8B-Experimental-1208-Instruct | d4b778ea-ae70-437f-a295-772abc659027 | 0.0.1 | hfopenllm_v2/sethuiyer_Llama-3.1-8B-Experimental-1208-Instruct/1762652580.513202 | 1762652580.513203 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | sethuiyer/Llama-3.1-8B-Experimental-1208-Instruct | sethuiyer/Llama-3.1-8B-Experimental-1208-Instruct | sethuiyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6099981382731153}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | sethuiyer | sethuiyer/Llamaverse-3.1-8B-Instruct | f0a224c2-037a-4229-bb00-5d76d3974078 | 0.0.1 | hfopenllm_v2/sethuiyer_Llamaverse-3.1-8B-Instruct/1762652580.513652 | 1762652580.513653 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | sethuiyer/Llamaverse-3.1-8B-Instruct | sethuiyer/Llamaverse-3.1-8B-Instruct | sethuiyer | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6185410266980501}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V23 | f83b7584-0e52-4658-ae15-f295064b9111 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V23/1762652579.904932 | 1762652579.9049332 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V23 | T145/ZEUS-8B-V23 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7621222799948582}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V12 | 1ab70352-9bda-47c8-8bdf-90860934cfc7 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V12/1762652579.901004 | 1762652579.901004 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V12 | T145/ZEUS-8B-V12 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.781556270695089}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V2-abliterated | 926fb6ed-0750-4d04-8e3c-da470e236db2 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V2-abliterated/1762652579.9039848 | 1762652579.903986 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V2-abliterated | T145/ZEUS-8B-V2-abliterated | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7895495064207414}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V24 | 51368b21-1b48-4c07-9b09-8cae0786200b | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V24/1762652579.905136 | 1762652579.9051368 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V24 | T145/ZEUS-8B-V24 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5999813827311533}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V21 | 380a44ec-387a-4f34-92c2-18fc7a8d5ce0 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V21/1762652579.904516 | 1762652579.904516 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V21 | T145/ZEUS-8B-V21 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3785145635801894}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V30 | 839ff423-8c5c-4fab-aecf-b535ee06af36 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V30/1762652579.907134 | 1762652579.907138 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V30 | T145/ZEUS-8B-V30 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7435626360279614}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/KRONOS-8B-V6 | 78813c35-3eaa-4ae6-9099-bf79efb8b0df | 0.0.1 | hfopenllm_v2/T145_KRONOS-8B-V6/1762652579.898935 | 1762652579.898936 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/KRONOS-8B-V6 | T145/KRONOS-8B-V6 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7022467054083166}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V9 | f5876dc1-b769-431f-84fe-365d2457902e | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V9/1762652579.908509 | 1762652579.90851 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V9 | T145/ZEUS-8B-V9 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5551436854213487}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V27 | bf31323b-bfb5-464a-b343-0605dafb5a60 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V27/1762652579.9057322 | 1762652579.905733 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V27 | T145/ZEUS-8B-V27 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.654361538495636}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/Meta-Llama-3.1-8B-Instruct-TIES | 15b92d44-3d68-4c6a-bddd-5676ebda2e10 | 0.0.1 | hfopenllm_v2/T145_Meta-Llama-3.1-8B-Instruct-TIES/1762652579.900369 | 1762652579.900369 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/Meta-Llama-3.1-8B-Instruct-TIES | T145/Meta-Llama-3.1-8B-Instruct-TIES | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5423542866261519}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V18 | e5d250e7-8d0a-48b5-aaad-3d1da02eab00 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V18/1762652579.903114 | 1762652579.903115 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V18 | T145/ZEUS-8B-V18 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7834046995305788}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V13-abliterated | 7c39d06a-dafe-40a7-b5a1-dca14dcadff2 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V13-abliterated/1762652579.901429 | 1762652579.9014301 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V13-abliterated | T145/ZEUS-8B-V13-abliterated | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7877509452696623}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V11 | a6eedf29-9ec8-4b03-a8f5-c9c4e2bda688 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V11/1762652579.900793 | 1762652579.900793 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V11 | T145/ZEUS-8B-V11 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8099575792231279}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V17 | 0368a3ba-e461-45d1-a037-3b9160a8efbb | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V17/1762652579.902262 | 1762652579.902263 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V17 | T145/ZEUS-8B-V17 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7940708431406447}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V6 | 09670c05-9463-479f-89e3-5029fd5d7ee7 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V6/1762652579.9077919 | 1762652579.9077928 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V6 | T145/ZEUS-8B-V6 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7837792612490415}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V10 | 464bae3d-bd06-4264-a939-59ab8e562ca6 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V10/1762652579.900583 | 1762652579.900584 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V10 | T145/ZEUS-8B-V10 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7706651684197928}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V25 | 52b41117-c308-4e8c-9c61-ce8e4faf778f | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V25/1762652579.905337 | 1762652579.905338 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V25 | T145/ZEUS-8B-V25 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.33202790817253774}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/KRONOS-8B-V3 | fc5613f1-09bc-4b82-89f4-4ee671cad5bf | 0.0.1 | hfopenllm_v2/T145_KRONOS-8B-V3/1762652579.8980181 | 1762652579.898019 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/KRONOS-8B-V3 | T145/KRONOS-8B-V3 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5474751437297483}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V26 | 8ae81cea-b179-4025-916a-9bc73755de82 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V26/1762652579.905539 | 1762652579.905539 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V26 | T145/ZEUS-8B-V26 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6707979272774018}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/KRONOS-8B-V1-P1 | 5bedfdac-2976-4a21-9ae2-a5b5b06e1e14 | 0.0.1 | hfopenllm_v2/T145_KRONOS-8B-V1-P1/1762652579.897121 | 1762652579.8971221 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/KRONOS-8B-V1-P1 | T145/KRONOS-8B-V1-P1 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7849783020164276}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V8 | c0035841-a312-493e-9c44-a75133e894d1 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V8/1762652579.908298 | 1762652579.908299 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V8 | T145/ZEUS-8B-V8 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7913979352562313}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V13 | 10823e50-9478-4a8a-83cf-5169a0bc1f1f | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V13/1762652579.90121 | 1762652579.9012108 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V13 | T145/ZEUS-8B-V13 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7904238531540756}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V2-ORPO | 588b0fce-37cd-41f1-8eaa-50383cdc0f00 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V2-ORPO/1762652579.903775 | 1762652579.903776 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V2-ORPO | T145/ZEUS-8B-V2-ORPO | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7186830941900824}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
HF Open LLM v2 | T145 | T145/KRONOS-8B-V4 | af8665b4-d9be-4243-9c8d-0b43e7abd540 | 0.0.1 | hfopenllm_v2/T145_KRONOS-8B-V4/1762652579.898447 | 1762652579.898448 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/KRONOS-8B-V4 | T145/KRONOS-8B-V4 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7889499860370484}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V17-abliterated-V2 | 3344d19c-c79b-48b3-be5b-f5f27d6920ce | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V17-abliterated-V2/1762652579.902674 | 1762652579.902674 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V17-abliterated-V2 | T145/ZEUS-8B-V17-abliterated-V2 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6532123654126606}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/Llama-3.1-8B-Instruct-Zeus | 38e620aa-c577-4b14-bebd-e98ebcbe48b2 | 0.0.1 | hfopenllm_v2/T145_Llama-3.1-8B-Instruct-Zeus/1762652579.899903 | 1762652579.899904 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/Llama-3.1-8B-Instruct-Zeus | T145/Llama-3.1-8B-Instruct-Zeus | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7941207108250552}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V3 | 6b8fca40-f44b-45a0-bd5b-04b2fa2067a2 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V3/1762652579.906709 | 1762652579.90671 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V3 | T145/ZEUS-8B-V3 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7886751596874072}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V7 | c6a9173a-bacc-40bd-9572-239f9901e065 | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V7/1762652579.908076 | 1762652579.908077 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V7 | T145/ZEUS-8B-V7 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7786085364610345}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V15 | 3e1be4f3-478f-4061-9856-f1beb0a749de | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V15/1762652579.901858 | 1762652579.901859 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V15 | T145/ZEUS-8B-V15 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.701272623306161}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
HF Open LLM v2 | T145 | T145/KRONOS-8B-V9 | 6fbb6156-196d-4523-900e-35316100d3b9 | 0.0.1 | hfopenllm_v2/T145_KRONOS-8B-V9/1762652579.8996658 | 1762652579.899667 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/KRONOS-8B-V9 | T145/KRONOS-8B-V9 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7855778224001206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/KRONOS-8B-V7 | 1358fee5-3874-4997-b1f0-6e93c6c5e9c0 | 0.0.1 | hfopenllm_v2/T145_KRONOS-8B-V7/1762652579.899169 | 1762652579.8991702 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/KRONOS-8B-V7 | T145/KRONOS-8B-V7 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3529102780622083}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V28 | e31561ff-779a-4ebe-b6fe-686b2895c53b | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V28/1762652579.905931 | 1762652579.905931 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V28 | T145/ZEUS-8B-V28 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.635252241829457}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | T145 | T145/ZEUS-8B-V17-abliterated | 35f89ab6-c6c9-41cd-9296-af4921490c3f | 0.0.1 | hfopenllm_v2/T145_ZEUS-8B-V17-abliterated/1762652579.902467 | 1762652579.9024682 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/ZEUS-8B-V17-abliterated | T145/ZEUS-8B-V17-abliterated | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7576009432749549}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.594} |
HF Open LLM v2 | T145 | T145/KRONOS-8B-V1-P2 | a5d0fc39-cac5-409f-8375-636ef97fba8c | 0.0.1 | hfopenllm_v2/T145_KRONOS-8B-V1-P2/1762652579.897378 | 1762652579.8973792 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | T145/KRONOS-8B-V1-P2 | T145/KRONOS-8B-V1-P2 | T145 | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6724213974476612}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.