_leaderboard
stringclasses
1 value
_developer
stringclasses
559 values
_model
stringlengths
9
102
_uuid
stringlengths
36
36
schema_version
stringclasses
1 value
evaluation_id
stringlengths
35
133
retrieved_timestamp
stringlengths
13
18
source_data
stringclasses
1 value
evaluation_source_name
stringclasses
1 value
evaluation_source_type
stringclasses
1 value
source_organization_name
stringclasses
1 value
source_organization_url
null
source_organization_logo_url
null
evaluator_relationship
stringclasses
1 value
model_name
stringlengths
4
102
model_id
stringlengths
9
102
model_developer
stringclasses
559 values
model_inference_platform
stringclasses
1 value
evaluation_results
stringlengths
1.35k
1.41k
additional_details
stringclasses
660 values
HF Open LLM v2
icefog72
icefog72/Ice0.62-18.01-RP
0c5bb530-f59b-4097-8a79-9e4f524385a2
0.0.1
hfopenllm_v2/icefog72_Ice0.62-18.01-RP/1762652580.21767
1762652580.217671
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.62-18.01-RP
icefog72/Ice0.62-18.01-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.536733644507684}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/IceSakeV8RP-7b
dbeb9a8a-53c5-472b-a4b1-1aa0582f8486
0.0.1
hfopenllm_v2/icefog72_IceSakeV8RP-7b/1762652580.2249868
1762652580.224988
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/IceSakeV8RP-7b
icefog72/IceSakeV8RP-7b
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6085741388404988}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.52.1-16.01-RP
6bfbd9d6-b376-4169-8e6a-2c3210040e97
0.0.1
hfopenllm_v2/icefog72_Ice0.52.1-16.01-RP/1762652580.21567
1762652580.215671
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.52.1-16.01-RP
icefog72/Ice0.52.1-16.01-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45492626231731803}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.27-06.11-RP
314c9c7e-0c13-4f6b-be25-d2a2cbc25e9b
0.0.1
hfopenllm_v2/icefog72_Ice0.27-06.11-RP/1762652580.211702
1762652580.211702
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.27-06.11-RP
icefog72/Ice0.27-06.11-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49182059158588104}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.51-16.01-RP
7a137ac4-8445-4c1a-9203-abc5f4131213
0.0.1
hfopenllm_v2/icefog72_Ice0.51-16.01-RP/1762652580.214901
1762652580.214902
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.51-16.01-RP
icefog72/Ice0.51-16.01-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4430610779398662}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/IceDrunkCherryRP-7b
c0e3f4ee-52dc-45c3-844a-8cc4e4520f24
0.0.1
hfopenllm_v2/icefog72_IceDrunkCherryRP-7b/1762652580.222923
1762652580.222924
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/IceDrunkCherryRP-7b
icefog72/IceDrunkCherryRP-7b
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48982255969715904}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.16-02.10-RP
824cb85d-e7a0-421a-994b-c0b178ab8e56
0.0.1
hfopenllm_v2/icefog72_Ice0.16-02.10-RP/1762652580.211284
1762652580.211284
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.16-02.10-RP
icefog72/Ice0.16-02.10-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5069083365470286}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.78-02.02-RP
ec943fa1-b138-46e8-b1ae-c9a476c73ed1
0.0.1
hfopenllm_v2/icefog72_Ice0.78-02.02-RP/1762652580.221266
1762652580.221267
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.78-02.02-RP
icefog72/Ice0.78-02.02-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.405292401937969}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/IceTea21EnergyDrinkRPV13-DPOv3.5
f4d3a112-d529-48f8-a99e-85e9eb02e0c1
0.0.1
hfopenllm_v2/icefog72_IceTea21EnergyDrinkRPV13-DPOv3.5/1762652580.2254012
1762652580.225402
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/IceTea21EnergyDrinkRPV13-DPOv3.5
icefog72/IceTea21EnergyDrinkRPV13-DPOv3.5
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48709978412833504}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/IceSakeV4RP-7b
93b5850f-74d0-45cd-977e-5bf6e4dc5d8d
0.0.1
hfopenllm_v2/icefog72_IceSakeV4RP-7b/1762652580.224551
1762652580.224552
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/IceSakeV4RP-7b
icefog72/IceSakeV4RP-7b
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4634192830578421}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.51.1-16.01-RP
859a9706-f73b-4426-9c5a-052625d62f5b
0.0.1
hfopenllm_v2/icefog72_Ice0.51.1-16.01-RP/1762652580.215148
1762652580.2151492
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.51.1-16.01-RP
icefog72/Ice0.51.1-16.01-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4573243438520902}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.70-25.01-RP
e109acd0-c7e3-4a9f-8e06-c428b95acc83
0.0.1
hfopenllm_v2/icefog72_Ice0.70-25.01-RP/1762652580.2196732
1762652580.219674
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.70-25.01-RP
icefog72/Ice0.70-25.01-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.549797869652522}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.73-01.02-RP
ba7bf09f-b7a1-4fd4-b262-4929a81da34a
0.0.1
hfopenllm_v2/icefog72_Ice0.73-01.02-RP/1762652580.220075
1762652580.220076
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.73-01.02-RP
icefog72/Ice0.73-01.02-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.529164838184905}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.31-08.11-RP
1fc072c6-ad31-4151-8420-7402b565510d
0.0.1
hfopenllm_v2/icefog72_Ice0.31-08.11-RP/1762652580.212094
1762652580.212095
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.31-08.11-RP
icefog72/Ice0.31-08.11-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5145768782386291}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.39-19.11-RP
780c711f-774b-499e-881e-25dba76273a1
0.0.1
hfopenllm_v2/icefog72_Ice0.39-19.11-RP/1762652580.2133162
1762652580.2133162
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.39-19.11-RP
icefog72/Ice0.39-19.11-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47565902915375646}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.17-03.10-RP
2faf039c-9c8e-46db-8472-6b741c451bf1
0.0.1
hfopenllm_v2/icefog72_Ice0.17-03.10-RP/1762652580.211494
1762652580.211495
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.17-03.10-RP
icefog72/Ice0.17-03.10-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5123538876846767}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.65-25.01-RP
fa5d2148-c45b-4266-a6a0-11b471273f75
0.0.1
hfopenllm_v2/icefog72_Ice0.65-25.01-RP/1762652580.2184708
1762652580.218472
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.65-25.01-RP
icefog72/Ice0.65-25.01-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5029366525264077}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.77-02.02-RP
0eebefc6-138f-4af5-a8b6-a35c798a38cb
0.0.1
hfopenllm_v2/icefog72_Ice0.77-02.02-RP/1762652580.221007
1762652580.2210078
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.77-02.02-RP
icefog72/Ice0.77-02.02-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5309633993359841}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/IceTea21EnergyDrinkRPV13-DPOv3
4b4a9630-c942-445e-b396-4a988d489aa7
0.0.1
hfopenllm_v2/icefog72_IceTea21EnergyDrinkRPV13-DPOv3/1762652580.225198
1762652580.2251992
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/IceTea21EnergyDrinkRPV13-DPOv3
icefog72/IceTea21EnergyDrinkRPV13-DPOv3
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5263423272472595}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.37-18.11-RP
774c0461-5e81-436a-9347-7a4cc15ca019
0.0.1
hfopenllm_v2/icefog72_Ice0.37-18.11-RP/1762652580.212915
1762652580.212916
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.37-18.11-RP
icefog72/Ice0.37-18.11-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4972162750391184}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.70.1-01.02-RP
ee088f70-5734-4951-8bc0-e0579a053fd2
0.0.1
hfopenllm_v2/icefog72_Ice0.70.1-01.02-RP/1762652580.219877
1762652580.219877
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.70.1-01.02-RP
icefog72/Ice0.70.1-01.02-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5069582042314393}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/IceDrinkByFrankensteinV3RP
39325b65-ad12-44ef-a1bf-ffe9e870ced8
0.0.1
hfopenllm_v2/icefog72_IceDrinkByFrankensteinV3RP/1762652580.222236
1762652580.222236
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/IceDrinkByFrankensteinV3RP
icefog72/IceDrinkByFrankensteinV3RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4974911013887596}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.62.1-24.01-RP
26ba869e-ae3b-44ef-a215-f94e4e4cb1fc
0.0.1
hfopenllm_v2/icefog72_Ice0.62.1-24.01-RP/1762652580.2178729
1762652580.2178729
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.62.1-24.01-RP
icefog72/Ice0.62.1-24.01-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5181740005407873}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.40-20.11-RP
5220bee5-74d3-4730-9fee-4ca488e1a37e
0.0.1
hfopenllm_v2/icefog72_Ice0.40-20.11-RP/1762652580.2136111
1762652580.213614
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.40-20.11-RP
icefog72/Ice0.40-20.11-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4762585495374495}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/IceDrinkNameGoesHereRP-7b-Model_Stock
b0aaf6e9-ffe3-4de9-b3f5-c33d52b59ed2
0.0.1
hfopenllm_v2/icefog72_IceDrinkNameGoesHereRP-7b-Model_Stock/1762652580.2224698
1762652580.2224698
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/IceDrinkNameGoesHereRP-7b-Model_Stock
icefog72/IceDrinkNameGoesHereRP-7b-Model_Stock
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49684171332065585}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.76-02.02-RP
701743bb-1ddf-4810-824a-38959d4a0e02
0.0.1
hfopenllm_v2/icefog72_Ice0.76-02.02-RP/1762652580.220735
1762652580.220737
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.76-02.02-RP
icefog72/Ice0.76-02.02-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45290274250100837}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.60.1-18.01-RP
8a14ed64-1408-469e-ab8d-05c897904d20
0.0.1
hfopenllm_v2/icefog72_Ice0.60.1-18.01-RP/1762652580.217258
1762652580.217259
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.60.1-18.01-RP
icefog72/Ice0.60.1-18.01-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5187735209244804}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.80-03.02-RP
847b4e14-a07c-45ed-b2eb-ecea0f80147b
0.0.1
hfopenllm_v2/icefog72_Ice0.80-03.02-RP/1762652580.2214909
1762652580.221492
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.80-03.02-RP
icefog72/Ice0.80-03.02-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5516462984880118}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.54-17.01-RP
94d01e56-d7d5-4680-b577-ebcc0198ca0c
0.0.1
hfopenllm_v2/icefog72_Ice0.54-17.01-RP/1762652580.2162719
1762652580.2162728
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.54-17.01-RP
icefog72/Ice0.54-17.01-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4378903531518593}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/IceNalyvkaRP-7b
95dd235d-6930-48fd-8594-5acb0110be29
0.0.1
hfopenllm_v2/icefog72_IceNalyvkaRP-7b/1762652580.224114
1762652580.224115
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/IceNalyvkaRP-7b
icefog72/IceNalyvkaRP-7b
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.549797869652522}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
icefog72
icefog72/Ice0.34b-14.11-RP
ed2a47c3-06c7-451b-94cd-8cd42be2ca9c
0.0.1
hfopenllm_v2/icefog72_Ice0.34b-14.11-RP/1762652580.2124958
1762652580.212497
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
icefog72/Ice0.34b-14.11-RP
icefog72/Ice0.34b-14.11-RP
icefog72
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47620868185303883}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 7.242}
HF Open LLM v2
bluuwhale
bluuwhale/L3-SthenoMaid-8B-V1
44dd13bc-56f0-4dd1-90d0-bb411239109a
0.0.1
hfopenllm_v2/bluuwhale_L3-SthenoMaid-8B-V1/1762652580.035146
1762652580.035147
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
bluuwhale/L3-SthenoMaid-8B-V1
bluuwhale/L3-SthenoMaid-8B-V1
bluuwhale
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7344700949037443}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-Orochi-v2-Experiment
db8eedcc-1dcf-47af-9c2b-a72da97146ca
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-Orochi-v2-Experiment/1762652579.5591779
1762652579.559179
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-Orochi-v2-Experiment
DoppelReflEx/MN-12B-Mimicore-Orochi-v2-Experiment
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.2842413684579139}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-Nocturne
6a21892f-1d11-4c59-8894-8800822b2e72
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-Nocturne/1762652579.558723
1762652579.5587242
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-Nocturne
DoppelReflEx/MN-12B-Mimicore-Nocturne
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3956502081144696}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-4
d7303703-f33e-430b-813d-998c95dbdb67
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-WhiteSnake-v2-Experiment-4/1762652579.560668
1762652579.560668
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-4
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-4
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42405151664250856}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-LilithFrame-Experiment-2
630c100f-c88d-42a7-9614-bd9a958eab2b
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-LilithFrame-Experiment-2/1762652579.5578592
1762652579.5578601
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-LilithFrame-Experiment-2
DoppelReflEx/MN-12B-LilithFrame-Experiment-2
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4299469851106176}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/L3-8B-WolfCore
c6771d5c-acaf-4b17-96b4-abf3b75bc68f
0.0.1
hfopenllm_v2/DoppelReflEx_L3-8B-WolfCore/1762652579.556399
1762652579.5564
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/L3-8B-WolfCore
DoppelReflEx/L3-8B-WolfCore
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4021950646506824}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MiniusLight-24B-v1b-test
2917ef74-c8cb-4255-8bda-76280fbe7c64
0.0.1
hfopenllm_v2/DoppelReflEx_MiniusLight-24B-v1b-test/1762652579.561931
1762652579.561932
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MiniusLight-24B-v1b-test
DoppelReflEx/MiniusLight-24B-v1b-test
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.37911408396388246}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 23.572}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-1
aa2478d9-59bd-458b-abee-5669aa6280df
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-WhiteSnake-v2-Experiment-1/1762652579.5600362
1762652579.5600362
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-1
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-1
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.39090391272933595}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Unleashed-Twilight
00f0fe96-4a06-46e7-88d8-368b86bcdb06
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Unleashed-Twilight/1762652579.560919
1762652579.56092
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Unleashed-Twilight
DoppelReflEx/MN-12B-Unleashed-Twilight
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3505121965274361}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-WhiteSnake
8aa34df4-8347-4f2d-98a0-7ec58bd62e43
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-WhiteSnake/1762652579.55982
1762652579.5598211
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-WhiteSnake
DoppelReflEx/MN-12B-Mimicore-WhiteSnake
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44376033369238066}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-2
66bd7a21-6f85-49b5-bc01-3f52ed8d1c64
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-WhiteSnake-v2-Experiment-2/1762652579.560246
1762652579.560246
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-2
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-2
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.31239333856389934}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-Orochi
f1bfef73-3586-4f9d-80ca-71b0fb00aadd
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-Orochi/1762652579.558937
1762652579.558938
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-Orochi
DoppelReflEx/MN-12B-Mimicore-Orochi
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4620451513096362}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/L3-8B-R1-WolfCore-V1.5-test
3c4058cd-238b-4b01-870d-8693f5ce1b8f
0.0.1
hfopenllm_v2/DoppelReflEx_L3-8B-R1-WolfCore-V1.5-test/1762652579.556192
1762652579.556193
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/L3-8B-R1-WolfCore-V1.5-test
DoppelReflEx/L3-8B-R1-WolfCore-V1.5-test
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3955006050612375}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/L3-8B-R1-WolfCore
6d8d63c0-ad69-4224-8250-b1664f6abbcf
0.0.1
hfopenllm_v2/DoppelReflEx_L3-8B-R1-WolfCore/1762652579.555949
1762652579.5559502
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/L3-8B-R1-WolfCore
DoppelReflEx/L3-8B-R1-WolfCore
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3775404814780339}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Kakigori
2f19082b-8377-4f63-8c5f-1aa25071a240
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Kakigori/1762652579.5572648
1762652579.557266
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Kakigori
DoppelReflEx/MN-12B-Kakigori
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.359329911302012}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-Orochi-v4-Experiment
e4e71999-6f83-4745-8a9d-66e711e39ac3
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-Orochi-v4-Experiment/1762652579.559606
1762652579.559606
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-Orochi-v4-Experiment
DoppelReflEx/MN-12B-Mimicore-Orochi-v4-Experiment
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4320702402957486}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-LilithFrame-Experiment-3
37292ca7-9e82-4c80-bc6e-bc7e1be7a95e
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-LilithFrame-Experiment-3/1762652579.558079
1762652579.558079
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-LilithFrame-Experiment-3
DoppelReflEx/MN-12B-LilithFrame-Experiment-3
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4127858526487498}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-Orochi-v3-Experiment
8198ab16-4a8b-4da9-8e8a-d1e3beb02839
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-Orochi-v3-Experiment/1762652579.559391
1762652579.559392
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-Orochi-v3-Experiment
DoppelReflEx/MN-12B-Mimicore-Orochi-v3-Experiment
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4101628124487471}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-GreenSnake
9b9eb072-4120-4a6a-a565-27136e617f10
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-GreenSnake/1762652579.5585039
1762652579.558505
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-GreenSnake
DoppelReflEx/MN-12B-Mimicore-GreenSnake
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47800724300411795}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MiniusLight-24B
2ec36e2e-0fba-4c6a-b9d0-fe57e7d708ef
0.0.1
hfopenllm_v2/DoppelReflEx_MiniusLight-24B/1762652579.561418
1762652579.561419
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MiniusLight-24B
DoppelReflEx/MiniusLight-24B
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.25766410900854175}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 23.572}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-3
1a3eefa6-7b3d-4541-93b0-8fe86f6bf038
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-Mimicore-WhiteSnake-v2-Experiment-3/1762652579.56046
1762652579.560461
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-3
DoppelReflEx/MN-12B-Mimicore-WhiteSnake-v2-Experiment-3
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4302218114602588}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MiniusLight-24B-test
e6031abf-1ae2-431c-8247-3124fff41d17
0.0.1
hfopenllm_v2/DoppelReflEx_MiniusLight-24B-test/1762652579.5616372
1762652579.5616379
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MiniusLight-24B-test
DoppelReflEx/MiniusLight-24B-test
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.03936776641533354}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 23.572}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-FoxFrame-test
ef5bb4eb-0875-4cc5-8e27-b59ffbd2e477
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-FoxFrame-test/1762652579.556618
1762652579.556619
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-FoxFrame-test
DoppelReflEx/MN-12B-FoxFrame-test
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42220308780701876}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-FoxFrame3-test
35351894-ea9d-456b-ab9a-c98686948e6b
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-FoxFrame3-test/1762652579.557049
1762652579.5570502
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-FoxFrame3-test
DoppelReflEx/MN-12B-FoxFrame3-test
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43231957871780213}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MiniusLight-24B-v1d-test
af67712e-7436-4703-ac22-9878dd8e190a
0.0.1
hfopenllm_v2/DoppelReflEx_MiniusLight-24B-v1d-test/1762652579.5624058
1762652579.5624058
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MiniusLight-24B-v1d-test
DoppelReflEx/MiniusLight-24B-v1d-test
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.40324339419407174}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 23.572}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-FoxFrame2-test
e46698de-8b2d-4b3c-b482-8cc8a3665eac
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-FoxFrame2-test/1762652579.556837
1762652579.5568378
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-FoxFrame2-test
DoppelReflEx/MN-12B-FoxFrame2-test
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43189514931492884}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-LilithFrame
a04a8775-8b4d-4608-9692-47af9f7ed5a7
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-LilithFrame/1762652579.557468
1762652579.557469
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-LilithFrame
DoppelReflEx/MN-12B-LilithFrame
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4509545782966972}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "float16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-LilithFrame
3d68e2fb-06cc-43b9-830b-f1cd02f12166
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-LilithFrame/1762652579.557674
1762652579.5576751
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-LilithFrame
DoppelReflEx/MN-12B-LilithFrame
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43604192431636946}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-LilithFrame-Experiment-4
ecc18f9c-c495-4ae6-8fd8-b2f84fb453ac
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-LilithFrame-Experiment-4/1762652579.5582879
1762652579.5582888
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-LilithFrame-Experiment-4
DoppelReflEx/MN-12B-LilithFrame-Experiment-4
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3981480250180632}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MiniusLight-24B-v1c-test
23a21492-0897-44b4-a046-cf93fa8c2a64
0.0.1
hfopenllm_v2/DoppelReflEx_MiniusLight-24B-v1c-test/1762652579.562173
1762652579.5621738
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MiniusLight-24B-v1c-test
DoppelReflEx/MiniusLight-24B-v1c-test
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.37858881102142317}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 23.572}
HF Open LLM v2
DoppelReflEx
DoppelReflEx/MN-12B-WolFrame
3bb96e7a-6c09-4b9e-8f2b-0b525c2ebeb3
0.0.1
hfopenllm_v2/DoppelReflEx_MN-12B-WolFrame/1762652579.5611808
1762652579.561182
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
DoppelReflEx/MN-12B-WolFrame
DoppelReflEx/MN-12B-WolFrame
DoppelReflEx
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4397387819873491}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "MistralForCausalLM", "params_billions": 12.248}
HF Open LLM v2
sethuiyer
sethuiyer/Llamazing-3.1-8B-Instruct
9065a7df-dab7-4e3b-bbc5-01f2908c37b3
0.0.1
hfopenllm_v2/sethuiyer_Llamazing-3.1-8B-Instruct/1762652580.513854
1762652580.513855
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
sethuiyer/Llamazing-3.1-8B-Instruct
sethuiyer/Llamazing-3.1-8B-Instruct
sethuiyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5711301568726534}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
sethuiyer
sethuiyer/Llama-3.1-8B-Experimental-1206-Instruct
49334550-08eb-49a2-9cea-f90f22533ab1
0.0.1
hfopenllm_v2/sethuiyer_Llama-3.1-8B-Experimental-1206-Instruct/1762652580.512954
1762652580.512954
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
sethuiyer/Llama-3.1-8B-Experimental-1206-Instruct
sethuiyer/Llama-3.1-8B-Experimental-1206-Instruct
sethuiyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6967014189018471}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
sethuiyer
sethuiyer/Llama-3.1-8B-Experimental-1208-Instruct
d4b778ea-ae70-437f-a295-772abc659027
0.0.1
hfopenllm_v2/sethuiyer_Llama-3.1-8B-Experimental-1208-Instruct/1762652580.513202
1762652580.513203
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
sethuiyer/Llama-3.1-8B-Experimental-1208-Instruct
sethuiyer/Llama-3.1-8B-Experimental-1208-Instruct
sethuiyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6099981382731153}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
sethuiyer
sethuiyer/Llamaverse-3.1-8B-Instruct
f0a224c2-037a-4229-bb00-5d76d3974078
0.0.1
hfopenllm_v2/sethuiyer_Llamaverse-3.1-8B-Instruct/1762652580.513652
1762652580.513653
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
sethuiyer/Llamaverse-3.1-8B-Instruct
sethuiyer/Llamaverse-3.1-8B-Instruct
sethuiyer
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6185410266980501}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V23
f83b7584-0e52-4658-ae15-f295064b9111
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V23/1762652579.904932
1762652579.9049332
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V23
T145/ZEUS-8B-V23
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7621222799948582}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V12
1ab70352-9bda-47c8-8bdf-90860934cfc7
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V12/1762652579.901004
1762652579.901004
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V12
T145/ZEUS-8B-V12
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.781556270695089}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V2-abliterated
926fb6ed-0750-4d04-8e3c-da470e236db2
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V2-abliterated/1762652579.9039848
1762652579.903986
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V2-abliterated
T145/ZEUS-8B-V2-abliterated
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7895495064207414}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V24
51368b21-1b48-4c07-9b09-8cae0786200b
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V24/1762652579.905136
1762652579.9051368
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V24
T145/ZEUS-8B-V24
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5999813827311533}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V21
380a44ec-387a-4f34-92c2-18fc7a8d5ce0
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V21/1762652579.904516
1762652579.904516
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V21
T145/ZEUS-8B-V21
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3785145635801894}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V30
839ff423-8c5c-4fab-aecf-b535ee06af36
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V30/1762652579.907134
1762652579.907138
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V30
T145/ZEUS-8B-V30
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7435626360279614}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/KRONOS-8B-V6
78813c35-3eaa-4ae6-9099-bf79efb8b0df
0.0.1
hfopenllm_v2/T145_KRONOS-8B-V6/1762652579.898935
1762652579.898936
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/KRONOS-8B-V6
T145/KRONOS-8B-V6
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7022467054083166}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V9
f5876dc1-b769-431f-84fe-365d2457902e
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V9/1762652579.908509
1762652579.90851
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V9
T145/ZEUS-8B-V9
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5551436854213487}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V27
bf31323b-bfb5-464a-b343-0605dafb5a60
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V27/1762652579.9057322
1762652579.905733
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V27
T145/ZEUS-8B-V27
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.654361538495636}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/Meta-Llama-3.1-8B-Instruct-TIES
15b92d44-3d68-4c6a-bddd-5676ebda2e10
0.0.1
hfopenllm_v2/T145_Meta-Llama-3.1-8B-Instruct-TIES/1762652579.900369
1762652579.900369
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/Meta-Llama-3.1-8B-Instruct-TIES
T145/Meta-Llama-3.1-8B-Instruct-TIES
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5423542866261519}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V18
e5d250e7-8d0a-48b5-aaad-3d1da02eab00
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V18/1762652579.903114
1762652579.903115
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V18
T145/ZEUS-8B-V18
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7834046995305788}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V13-abliterated
7c39d06a-dafe-40a7-b5a1-dca14dcadff2
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V13-abliterated/1762652579.901429
1762652579.9014301
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V13-abliterated
T145/ZEUS-8B-V13-abliterated
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7877509452696623}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V11
a6eedf29-9ec8-4b03-a8f5-c9c4e2bda688
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V11/1762652579.900793
1762652579.900793
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V11
T145/ZEUS-8B-V11
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.8099575792231279}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V17
0368a3ba-e461-45d1-a037-3b9160a8efbb
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V17/1762652579.902262
1762652579.902263
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V17
T145/ZEUS-8B-V17
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7940708431406447}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V6
09670c05-9463-479f-89e3-5029fd5d7ee7
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V6/1762652579.9077919
1762652579.9077928
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V6
T145/ZEUS-8B-V6
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7837792612490415}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V10
464bae3d-bd06-4264-a939-59ab8e562ca6
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V10/1762652579.900583
1762652579.900584
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V10
T145/ZEUS-8B-V10
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7706651684197928}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V25
52b41117-c308-4e8c-9c61-ce8e4faf778f
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V25/1762652579.905337
1762652579.905338
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V25
T145/ZEUS-8B-V25
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.33202790817253774}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/KRONOS-8B-V3
fc5613f1-09bc-4b82-89f4-4ee671cad5bf
0.0.1
hfopenllm_v2/T145_KRONOS-8B-V3/1762652579.8980181
1762652579.898019
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/KRONOS-8B-V3
T145/KRONOS-8B-V3
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5474751437297483}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V26
8ae81cea-b179-4025-916a-9bc73755de82
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V26/1762652579.905539
1762652579.905539
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V26
T145/ZEUS-8B-V26
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6707979272774018}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/KRONOS-8B-V1-P1
5bedfdac-2976-4a21-9ae2-a5b5b06e1e14
0.0.1
hfopenllm_v2/T145_KRONOS-8B-V1-P1/1762652579.897121
1762652579.8971221
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/KRONOS-8B-V1-P1
T145/KRONOS-8B-V1-P1
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7849783020164276}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V8
c0035841-a312-493e-9c44-a75133e894d1
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V8/1762652579.908298
1762652579.908299
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V8
T145/ZEUS-8B-V8
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7913979352562313}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V13
10823e50-9478-4a8a-83cf-5169a0bc1f1f
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V13/1762652579.90121
1762652579.9012108
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V13
T145/ZEUS-8B-V13
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7904238531540756}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V2-ORPO
588b0fce-37cd-41f1-8eaa-50383cdc0f00
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V2-ORPO/1762652579.903775
1762652579.903776
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V2-ORPO
T145/ZEUS-8B-V2-ORPO
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7186830941900824}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015}
HF Open LLM v2
T145
T145/KRONOS-8B-V4
af8665b4-d9be-4243-9c8d-0b43e7abd540
0.0.1
hfopenllm_v2/T145_KRONOS-8B-V4/1762652579.898447
1762652579.898448
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/KRONOS-8B-V4
T145/KRONOS-8B-V4
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7889499860370484}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V17-abliterated-V2
3344d19c-c79b-48b3-be5b-f5f27d6920ce
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V17-abliterated-V2/1762652579.902674
1762652579.902674
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V17-abliterated-V2
T145/ZEUS-8B-V17-abliterated-V2
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6532123654126606}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/Llama-3.1-8B-Instruct-Zeus
38e620aa-c577-4b14-bebd-e98ebcbe48b2
0.0.1
hfopenllm_v2/T145_Llama-3.1-8B-Instruct-Zeus/1762652579.899903
1762652579.899904
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/Llama-3.1-8B-Instruct-Zeus
T145/Llama-3.1-8B-Instruct-Zeus
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7941207108250552}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V3
6b8fca40-f44b-45a0-bd5b-04b2fa2067a2
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V3/1762652579.906709
1762652579.90671
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V3
T145/ZEUS-8B-V3
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7886751596874072}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V7
c6a9173a-bacc-40bd-9572-239f9901e065
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V7/1762652579.908076
1762652579.908077
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V7
T145/ZEUS-8B-V7
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7786085364610345}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V15
3e1be4f3-478f-4061-9856-f1beb0a749de
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V15/1762652579.901858
1762652579.901859
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V15
T145/ZEUS-8B-V15
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.701272623306161}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015}
HF Open LLM v2
T145
T145/KRONOS-8B-V9
6fbb6156-196d-4523-900e-35316100d3b9
0.0.1
hfopenllm_v2/T145_KRONOS-8B-V9/1762652579.8996658
1762652579.899667
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/KRONOS-8B-V9
T145/KRONOS-8B-V9
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7855778224001206}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/KRONOS-8B-V7
1358fee5-3874-4997-b1f0-6e93c6c5e9c0
0.0.1
hfopenllm_v2/T145_KRONOS-8B-V7/1762652579.899169
1762652579.8991702
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/KRONOS-8B-V7
T145/KRONOS-8B-V7
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.3529102780622083}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015}
HF Open LLM v2
T145
T145/ZEUS-8B-V28
e31561ff-779a-4ebe-b6fe-686b2895c53b
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V28/1762652579.905931
1762652579.905931
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V28
T145/ZEUS-8B-V28
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.635252241829457}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}
HF Open LLM v2
T145
T145/ZEUS-8B-V17-abliterated
35f89ab6-c6c9-41cd-9296-af4921490c3f
0.0.1
hfopenllm_v2/T145_ZEUS-8B-V17-abliterated/1762652579.902467
1762652579.9024682
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/ZEUS-8B-V17-abliterated
T145/ZEUS-8B-V17-abliterated
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.7576009432749549}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 7.594}
HF Open LLM v2
T145
T145/KRONOS-8B-V1-P2
a5d0fc39-cac5-409f-8375-636ef97fba8c
0.0.1
hfopenllm_v2/T145_KRONOS-8B-V1-P2/1762652579.897378
1762652579.8973792
["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"]
HF Open LLM v2
leaderboard
Hugging Face
null
null
third_party
T145/KRONOS-8B-V1-P2
T145/KRONOS-8B-V1-P2
T145
unknown
[{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.6724213974476612}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH...
{"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03}