_leaderboard stringclasses 1
value | _developer stringclasses 559
values | _model stringlengths 9 102 | _uuid stringlengths 36 36 | schema_version stringclasses 1
value | evaluation_id stringlengths 35 133 | retrieved_timestamp stringlengths 13 18 | source_data stringclasses 1
value | evaluation_source_name stringclasses 1
value | evaluation_source_type stringclasses 1
value | source_organization_name stringclasses 1
value | source_organization_url null | source_organization_logo_url null | evaluator_relationship stringclasses 1
value | model_name stringlengths 4 102 | model_id stringlengths 9 102 | model_developer stringclasses 559
values | model_inference_platform stringclasses 1
value | evaluation_results stringlengths 1.35k 1.41k | additional_details stringclasses 660
values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
HF Open LLM v2 | jaspionjader | jaspionjader/sof-3 | e5cd6a8b-88ed-4a0d-8584-889a4fde72a7 | 0.0.1 | hfopenllm_v2/jaspionjader_sof-3/1762652580.277219 | 1762652580.27722 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/sof-3 | jaspionjader/sof-3 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46369410940748323}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v24-8B | e6b62da0-ad6d-431c-8a0e-185c6eddf3da | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v24-8B/1762652580.233697 | 1762652580.2336981 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v24-8B | jaspionjader/Kosmos-EVAA-PRP-v24-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42587556572117535}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-13 | da5a3c32-371f-44e5-89a7-c9ba6e98664e | 0.0.1 | hfopenllm_v2/jaspionjader_bh-13/1762652580.248588 | 1762652580.248588 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-13 | jaspionjader/bh-13 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4697890486132351}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-17 | 787d8040-25c8-4893-b140-cf041260d767 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-17/1762652580.249204 | 1762652580.2492049 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-17 | jaspionjader/bh-17 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4721871301480073}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-51 | 6d544c96-53c9-43d1-9cb1-6077d7235fff | 0.0.1 | hfopenllm_v2/jaspionjader_bh-51/1762652580.2581341 | 1762652580.258135 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-51 | jaspionjader/bh-51 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4630447213393795}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-6 | e8dfd77c-e2c8-42ef-b341-5476411d038d | 0.0.1 | hfopenllm_v2/jaspionjader_bh-6/1762652580.260308 | 1762652580.260309 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-6 | jaspionjader/bh-6 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4620706392372239}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-20 | cfe4ab09-c772-4617-88b6-77e49553605b | 0.0.1 | hfopenllm_v2/jaspionjader_bh-20/1762652580.2503 | 1762652580.2503 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-20 | jaspionjader/bh-20 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4727367828472896}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/f-9-8b | 1373c279-13b7-46d3-94a4-7b47c9319f88 | 0.0.1 | hfopenllm_v2/jaspionjader_f-9-8b/1762652580.267217 | 1762652580.2672179 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/f-9-8b | jaspionjader/f-9-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4601723427173233}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v7-8B | 798c2f08-e10b-4115-bdd5-0d6053d03b60 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v7-8B/1762652580.242492 | 1762652580.242493 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v7-8B | jaspionjader/Kosmos-EVAA-v7-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4276741268722545}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-TSN-v20-8B | 4a60fea6-e0e8-497e-9b29-439e7641e77b | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-TSN-v20-8B/1762652580.236737 | 1762652580.236737 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-TSN-v20-8B | jaspionjader/Kosmos-EVAA-TSN-v20-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4423119545029411}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/fct-14-8b | 22c3022f-d538-4a4d-8d4b-05e915506451 | 0.0.1 | hfopenllm_v2/jaspionjader_fct-14-8b/1762652580.2674618 | 1762652580.267463 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/fct-14-8b | jaspionjader/fct-14-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4128612082607481}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-42 | 265e3cbb-484f-4cf7-8994-050f414ecf37 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-42/1762652580.25606 | 1762652580.2560608 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-42 | jaspionjader/bh-42 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4660423232578447}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-Franken-v38-8B | 8919b3ad-529c-4391-bec3-65b81dad97c3 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-Franken-v38-8B/1762652580.2324722 | 1762652580.2324731 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-Franken-v38-8B | jaspionjader/Kosmos-EVAA-Franken-v38-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4355676272290855}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-Elusive-VENN-8B | d3af54be-9d9a-4a4a-b03e-3468a801795e | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-Elusive-VENN-8B/1762652580.243592 | 1762652580.243593 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-Elusive-VENN-8B | jaspionjader/Kosmos-Elusive-VENN-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4232525255211727}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-20 | 1430e550-80ca-4f84-952f-b5b10fbca711 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-20/1762652580.273083 | 1762652580.273084 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-20 | jaspionjader/slu-20 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4393143525844759}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-37 | 19490f78-486d-4325-b31e-af8555c32ea9 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-37/1762652580.2544441 | 1762652580.254445 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-37 | jaspionjader/bh-37 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48797413086166924}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/kstc-9-8b | 5f36e182-fa70-41d9-9cc6-12367035fc76 | 0.0.1 | hfopenllm_v2/jaspionjader_kstc-9-8b/1762652580.27159 | 1762652580.27159 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/kstc-9-8b | jaspionjader/kstc-9-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4860758343417687}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-38 | 61e7c49e-abb9-4e38-ba3f-1018db104d83 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-38/1762652580.2548852 | 1762652580.2548869 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-38 | jaspionjader/bh-38 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46179581288758276}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-30 | 6d3a64df-5ebb-4cd8-bd6c-de799d185fe1 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-30/1762652580.252943 | 1762652580.2529438 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-30 | jaspionjader/bh-30 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46664184364153777}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-39 | 243e6b7b-a34f-44cd-b027-176f877ff8e7 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-39/1762652580.2552152 | 1762652580.2552161 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-39 | jaspionjader/bh-39 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45759917020173135}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-Elusive-VENN-Asymmetric-8B | e7cf15b2-0347-48a8-bf84-08e27b3688fd | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-Elusive-VENN-Asymmetric-8B/1762652580.243807 | 1762652580.243807 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-Elusive-VENN-Asymmetric-8B | jaspionjader/Kosmos-Elusive-VENN-Asymmetric-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4541771388803929}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v9-TitanFusion-Mix-8B | 69f3e2b2-8918-41a8-abc6-c84c3d674f94 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v9-TitanFusion-Mix-8B/1762652580.243146 | 1762652580.243147 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v9-TitanFusion-Mix-8B | jaspionjader/Kosmos-EVAA-v9-TitanFusion-Mix-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.428373382624769}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-40 | 56837896-11a6-458b-a17e-9540ab5ae66a | 0.0.1 | hfopenllm_v2/jaspionjader_bh-40/1762652580.2556531 | 1762652580.2556539 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-40 | jaspionjader/bh-40 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45357761849669986}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v6-8B | c0cc1ad5-9e53-45ac-becb-f8ce3e5ac631 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v6-8B/1762652580.242274 | 1762652580.242275 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v6-8B | jaspionjader/Kosmos-EVAA-v6-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4395891789341171}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v11-8B | 9a6b85d5-bb26-4832-915e-8b1ac90b0793 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v11-8B/1762652580.240909 | 1762652580.24091 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v11-8B | jaspionjader/Kosmos-EVAA-v11-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44263664853699297}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-60 | 16d14b95-fe8b-4e1f-94e1-65d966ba24d6 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-60/1762652580.2605288 | 1762652580.2605288 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-60 | jaspionjader/bh-60 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42070484093316846}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/f-1-8b | 91d65b2a-a96a-467b-9e5c-9efa28d7fd96 | 0.0.1 | hfopenllm_v2/jaspionjader_f-1-8b/1762652580.264415 | 1762652580.264416 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/f-1-8b | jaspionjader/f-1-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49826571275327247}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/ek-6 | a05ce252-928c-4482-95f7-f4c0fc2c7c10 | 0.0.1 | hfopenllm_v2/jaspionjader_ek-6/1762652580.2637498 | 1762652580.263751 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/ek-6 | jaspionjader/ek-6 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4642437621067656}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-3 | 37e59290-b4ea-4a44-bfb0-cdbe781c4d7f | 0.0.1 | hfopenllm_v2/jaspionjader_bh-3/1762652580.2527301 | 1762652580.2527308 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-3 | jaspionjader/bh-3 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4663670172918966}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-43 | 472b725a-2bd5-440a-9768-ba8db6fe6b34 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-43/1762652580.2562718 | 1762652580.2562718 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-43 | jaspionjader/bh-43 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45999725173650363}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Auro-Kosmos-EVAA-v2.1-8B | 4381d7ab-d19f-4fa0-a69a-978af28df8fa | 0.0.1 | hfopenllm_v2/jaspionjader_Auro-Kosmos-EVAA-v2.1-8B/1762652580.231263 | 1762652580.231264 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Auro-Kosmos-EVAA-v2.1-8B | jaspionjader/Auro-Kosmos-EVAA-v2.1-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4665919759571271}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v32-8B | f58f0ecc-a059-448d-a2f9-e36b601e2154 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v32-8B/1762652580.235436 | 1762652580.2354372 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v32-8B | jaspionjader/Kosmos-EVAA-PRP-v32-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4487315877427448}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-36 | cf85253f-0ecd-4943-a508-eab1e562a497 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-36/1762652580.275441 | 1762652580.275442 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-36 | jaspionjader/slu-36 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4518289250300314}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/kstc-8-8b | 85502cb7-db11-43ce-a3cf-f9329ecec2e1 | 0.0.1 | hfopenllm_v2/jaspionjader_kstc-8-8b/1762652580.271383 | 1762652580.271384 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/kstc-8-8b | jaspionjader/kstc-8-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.49097173278013445}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-34 | 6097086b-8c8b-493e-af1a-71146a2ed566 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-34/1762652580.253809 | 1762652580.25381 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-34 | jaspionjader/bh-34 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4623953332712758}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-mix-1 | 3a8a175f-5173-491b-9acf-87fe781f16df | 0.0.1 | hfopenllm_v2/jaspionjader_slu-mix-1/1762652580.276264 | 1762652580.276264 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-mix-1 | jaspionjader/slu-mix-1 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45689991444921696}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-Elusive-VENN-Aurora_faustus-8B | 8befbe9f-3ab2-4bc8-bd16-5badd2291d5d | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-Elusive-VENN-Aurora_faustus-8B/1762652580.244045 | 1762652580.244046 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-Elusive-VENN-Aurora_faustus-8B | jaspionjader/Kosmos-Elusive-VENN-Aurora_faustus-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4335441074127758}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/PRP-Kosmos-EVAA-light-8B | 72c9dcd4-ab00-4f36-a1e6-43e241c8b967 | 0.0.1 | hfopenllm_v2/jaspionjader_PRP-Kosmos-EVAA-light-8B/1762652580.2449658 | 1762652580.244967 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/PRP-Kosmos-EVAA-light-8B | jaspionjader/PRP-Kosmos-EVAA-light-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4321201079801593}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v33-8B | 2436838e-2b6a-4c1e-b8c2-ec505d9a4c34 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v33-8B/1762652580.23565 | 1762652580.235651 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v33-8B | jaspionjader/Kosmos-EVAA-PRP-v33-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4301719437758481}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-Franken-Immersive-v39-8B | f9e1901a-854d-4437-8d49-a6c47799f687 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-Franken-Immersive-v39-8B/1762652580.232267 | 1762652580.232268 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-Franken-Immersive-v39-8B | jaspionjader/Kosmos-EVAA-Franken-Immersive-v39-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43779061778303796}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-50 | 980887dd-2948-4e5f-b22c-3cc03057f493 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-50/1762652580.257925 | 1762652580.257926 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-50 | jaspionjader/bh-50 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47246195649764844}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/ek-7 | 23127691-ff90-433f-97d2-322e1191d821 | 0.0.1 | hfopenllm_v2/jaspionjader_ek-7/1762652580.264135 | 1762652580.2641358 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/ek-7 | jaspionjader/ek-7 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47670846686791046}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-v18-8B | aadb6262-4f31-4681-983c-0d19e8bbc5cd | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-v18-8B/1762652580.240138 | 1762652580.240139 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-v18-8B | jaspionjader/Kosmos-EVAA-gamma-v18-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43409376011205825}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bbb-2 | 155b7412-cc16-45c3-9261-acc9322a0dcc | 0.0.1 | hfopenllm_v2/jaspionjader_bbb-2/1762652580.2460952 | 1762652580.2460968 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bbb-2 | jaspionjader/bbb-2 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4077403511571519}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/gamma-Kosmos-EVAA-v2-8B | 67f972e1-4ebd-4b78-b740-fdc03ac88aac | 0.0.1 | hfopenllm_v2/jaspionjader_gamma-Kosmos-EVAA-v2-8B/1762652580.268805 | 1762652580.268806 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/gamma-Kosmos-EVAA-v2-8B | jaspionjader/gamma-Kosmos-EVAA-v2-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4232525255211727}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-27 | 68435a43-944b-4c66-979b-eb48f7a8e77a | 0.0.1 | hfopenllm_v2/jaspionjader_bh-27/1762652580.2520802 | 1762652580.252081 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-27 | jaspionjader/bh-27 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4818791916559174}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-18 | 3f3eeca1-d401-436e-b7e6-5fa82c099270 | 0.0.1 | hfopenllm_v2/jaspionjader_test-18/1762652580.2796118 | 1762652580.279613 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-18 | jaspionjader/test-18 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43916474953124374}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bbb-5 | 8c0a66fb-c87d-489d-b071-b4a599562ead | 0.0.1 | hfopenllm_v2/jaspionjader_bbb-5/1762652580.2468202 | 1762652580.2468212 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bbb-5 | jaspionjader/bbb-5 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4702888336281067}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-light-8B | 64c07a98-4f3f-49f7-99de-9963dcfedeba | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-light-8B/1762652580.237838 | 1762652580.2378392 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-light-8B | jaspionjader/Kosmos-EVAA-gamma-light-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45809895521660304}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-light-8B | 0d2e1c3f-8ee6-44b0-912a-452e2a5a6da7 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-light-8B/1762652580.233289 | 1762652580.23329 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-light-8B | jaspionjader/Kosmos-EVAA-PRP-light-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.38238651223198894}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-19 | 81914fd7-1410-4b80-9be9-6ebfbb664613 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-19/1762652580.249828 | 1762652580.249829 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-19 | jaspionjader/bh-19 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45842364925065493}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/fr-3-8b | 8bdd1aba-81e4-44d1-acfd-6efeaf391ac8 | 0.0.1 | hfopenllm_v2/jaspionjader_fr-3-8b/1762652580.268359 | 1762652580.26836 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/fr-3-8b | jaspionjader/fr-3-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4325700253106203}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v30-8B | 320c581d-f667-4dab-a32c-bb9f2621e84d | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v30-8B/1762652580.2349901 | 1762652580.234991 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v30-8B | jaspionjader/Kosmos-EVAA-PRP-v30-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42947268802333366}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-46 | 6b3c3872-cd4d-4827-8651-6baa9d2423e7 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-46/1762652580.2567308 | 1762652580.256732 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-46 | jaspionjader/bh-46 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4727367828472896}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/dp-7-8b | 44d85302-1af8-48ef-aebe-a9512c5bc387 | 0.0.1 | hfopenllm_v2/jaspionjader_dp-7-8b/1762652580.2634509 | 1762652580.2634518 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/dp-7-8b | jaspionjader/dp-7-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44983089314130953}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/f-2-8b | c63fc798-cf74-4767-ba95-6353b6761bcc | 0.0.1 | hfopenllm_v2/jaspionjader_f-2-8b/1762652580.264705 | 1762652580.2647061 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/f-2-8b | jaspionjader/f-2-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48237897667078905}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/sof-1 | b1f4196a-0050-4107-a97b-4e1bd6ece17b | 0.0.1 | hfopenllm_v2/jaspionjader_sof-1/1762652580.276484 | 1762652580.2764852 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/sof-1 | jaspionjader/sof-1 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4313709845432342}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v12-8B | 4bcdbab0-7220-40bb-832f-01003f59da0f | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v12-8B/1762652580.2411451 | 1762652580.241146 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v12-8B | jaspionjader/Kosmos-EVAA-v12-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43779061778303796}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-1 | 15ec7997-1333-43c6-869a-ce4589af56d1 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-1/1762652580.2475939 | 1762652580.247595 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-1 | jaspionjader/bh-1 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42842325030917966}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-22 | f3815ff9-c1bd-4706-a770-4c0b0e8c5d13 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-22/1762652580.250869 | 1762652580.25087 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-22 | jaspionjader/bh-22 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45999725173650363}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-35 | 7166192e-42b0-4990-8218-88bb38fd1bdb | 0.0.1 | hfopenllm_v2/jaspionjader_bh-35/1762652580.2540212 | 1762652580.254022 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-35 | jaspionjader/bh-35 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.47213726246359655}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-57 | fab7388c-87ed-4108-ba4d-e1621925f264 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-57/1762652580.259624 | 1762652580.259625 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-57 | jaspionjader/bh-57 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44051339335186196}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-15 | f197c7ce-c30a-49ad-bd6c-9571d3b25637 | 0.0.1 | hfopenllm_v2/jaspionjader_test-15/1762652580.278964 | 1762652580.278965 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-15 | jaspionjader/test-15 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4364918416468304}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-23 | f4db95ae-8e3d-45ed-9c53-3b30fde0cb3e | 0.0.1 | hfopenllm_v2/jaspionjader_bh-23/1762652580.2511601 | 1762652580.251161 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-23 | jaspionjader/bh-23 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46576749690820357}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v3-8B | 262a66ee-04e4-49d5-8bb2-efe0a93801ad | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v3-8B/1762652580.241601 | 1762652580.241602 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v3-8B | jaspionjader/Kosmos-EVAA-v3-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4410630460511443}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-v16-8B | 15deaa33-87a2-442e-9618-13f5ab6c299e | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-v16-8B/1762652580.2392871 | 1762652580.239288 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-v16-8B | jaspionjader/Kosmos-EVAA-gamma-v16-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4556510059974202}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-TSN-v22-8B | 6e20f902-8752-466c-b8d4-34787fb90fce | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-TSN-v22-8B/1762652580.2371762 | 1762652580.2371771 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-TSN-v22-8B | jaspionjader/Kosmos-EVAA-TSN-v22-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4673410993940522}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bbb-7 | a31fbd82-2e21-40e7-a73a-c6351c80bae7 | 0.0.1 | hfopenllm_v2/jaspionjader_bbb-7/1762652580.2473001 | 1762652580.247304 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bbb-7 | jaspionjader/bbb-7 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.48280340607366234}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-56 | 348c8f2b-807f-464b-832e-0049f8329b86 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-56/1762652580.2593641 | 1762652580.259365 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-56 | jaspionjader/bh-56 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45999725173650363}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-VENN-8B | e14cedfb-79a9-446a-ba16-64f378a47b4a | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-VENN-8B/1762652580.24428 | 1762652580.244281 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-VENN-8B | jaspionjader/Kosmos-VENN-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.433219413378724}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "float16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-gamma-v13-8B | 1fc6ca13-157c-4502-8724-be153afb6347 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-gamma-v13-8B/1762652580.238605 | 1762652580.238605 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-gamma-v13-8B | jaspionjader/Kosmos-EVAA-gamma-v13-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44286160720222345}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-9 | 956d92e9-51fb-4770-8687-6003f9594345 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-9/1762652580.262652 | 1762652580.2626529 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-9 | jaspionjader/bh-9 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4508548429278758}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/sof-10 | 03761253-711d-428d-a3bd-89974a50b490 | 0.0.1 | hfopenllm_v2/jaspionjader_sof-10/1762652580.276895 | 1762652580.276897 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/sof-10 | jaspionjader/sof-10 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46484328249045864}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-v8-8B | 388ef85a-db27-4851-9e6e-2002a75bc6c7 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-v8-8B/1762652580.242712 | 1762652580.242713 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-v8-8B | jaspionjader/Kosmos-EVAA-v8-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43834027048232027}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-5 | ec314c97-9bc0-4e14-9d57-d6204e699428 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-5/1762652580.2577002 | 1762652580.2577012 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-5 | jaspionjader/bh-5 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.46516797652451053}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-58 | a9c1b649-8850-43d1-b5db-feefd0b8d0b4 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-58/1762652580.259867 | 1762652580.259868 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-58 | jaspionjader/bh-58 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4630447213393795}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-10 | 79c255e5-8a6b-4afd-a03e-e35cbcbcc712 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-10/1762652580.271806 | 1762652580.271807 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-10 | jaspionjader/slu-10 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4359920566319587}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-8B | eb68e0e3-1e39-4779-bc99-4e1825d9c602 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-8B/1762652580.233048 | 1762652580.2330492 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-8B | jaspionjader/Kosmos-EVAA-PRP-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.34052092891306174}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-29 | fe231e36-6cc2-412c-b86e-0ba6ba9cc430 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-29/1762652580.274164 | 1762652580.274165 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-29 | jaspionjader/slu-29 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4430610779398662}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-33 | 2597a3df-0f30-43d1-b1b3-7a0baac07675 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-33/1762652580.274691 | 1762652580.274692 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-33 | jaspionjader/slu-33 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4457339858242796}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-44 | 60c18178-ff40-4e9d-9683-077cc2fa254e | 0.0.1 | hfopenllm_v2/jaspionjader_bh-44/1762652580.2565289 | 1762652580.2565298 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-44 | jaspionjader/bh-44 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4706135276621586}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Auro-Kosmos-EVAA-v2-8B | 0064f2f6-672e-478c-9184-e7fd32ad06b8 | 0.0.1 | hfopenllm_v2/jaspionjader_Auro-Kosmos-EVAA-v2-8B/1762652580.231028 | 1762652580.231029 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Auro-Kosmos-EVAA-v2-8B | jaspionjader/Auro-Kosmos-EVAA-v2-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4778077722664752}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/dp-6-8b | 5c61d4f5-25a0-4ffe-a9d2-2a33d8bbd717 | 0.0.1 | hfopenllm_v2/jaspionjader_dp-6-8b/1762652580.263117 | 1762652580.2631192 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/dp-6-8b | jaspionjader/dp-6-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4805804155197099}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-35 | 0d7698b6-de52-4781-831f-a3ca8b23dd72 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-35/1762652580.275198 | 1762652580.2752 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-35 | jaspionjader/slu-35 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42417673993891764}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v26-8B | 6705072a-5a46-49ae-925f-1cf7da1ea288 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v26-8B/1762652580.234126 | 1762652580.234127 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v26-8B | jaspionjader/Kosmos-EVAA-PRP-v26-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4413877400851962}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/f-5-8b | 4dd614dc-b68b-456c-ac55-f2221a479caa | 0.0.1 | hfopenllm_v2/jaspionjader_f-5-8b/1762652580.265783 | 1762652580.265785 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/f-5-8b | jaspionjader/f-5-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.5043606519590242}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-36 | 3a4f8c97-9f30-44b8-8f79-7f19f90a08d1 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-36/1762652580.2542279 | 1762652580.254229 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-36 | jaspionjader/bh-36 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4665919759571271}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/bh-52 | fd3c9666-09bf-4562-b49d-eea905469761 | 0.0.1 | hfopenllm_v2/jaspionjader_bh-52/1762652580.258348 | 1762652580.258349 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/bh-52 | jaspionjader/bh-52 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45362748618111054}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-12 | e49c9cc8-96ff-4a3c-b7b4-ea5562f41449 | 0.0.1 | hfopenllm_v2/jaspionjader_test-12/1762652580.278201 | 1762652580.278202 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-12 | jaspionjader/test-12 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4368165356808823}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-2 | 1950fba0-3a1b-4cbe-8fa5-9947ed8e4bad | 0.0.1 | hfopenllm_v2/jaspionjader_slu-2/1762652580.2728698 | 1762652580.272871 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-2 | jaspionjader/slu-2 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.40159554426698935}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-23 | f4b76351-e472-47a9-8011-6bf2e7e33a71 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-23/1762652580.27371 | 1762652580.2737112 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-23 | jaspionjader/slu-23 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44780737332499987}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/sof-6 | 0755b7f9-bdd7-4e2a-92da-6650934db265 | 0.0.1 | hfopenllm_v2/jaspionjader_sof-6/1762652580.277473 | 1762652580.2774742 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/sof-6 | jaspionjader/sof-6 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4353925362482657}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-PRP-v28-8B | e2aa230d-452e-42f0-a780-af255c62120e | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-PRP-v28-8B/1762652580.234553 | 1762652580.234553 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-PRP-v28-8B | jaspionjader/Kosmos-EVAA-PRP-v28-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.43659157701565177}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-17 | fea528ae-4015-4adf-bce0-f9775554cc5f | 0.0.1 | hfopenllm_v2/jaspionjader_slu-17/1762652580.272654 | 1762652580.272655 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-17 | jaspionjader/slu-17 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42167892303532406}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-17 | c9933c3d-98ab-4486-bd42-7c90f5ed3bd2 | 0.0.1 | hfopenllm_v2/jaspionjader_test-17/1762652580.279401 | 1762652580.279402 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-17 | jaspionjader/test-17 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.42674991245450955}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
HF Open LLM v2 | jaspionjader | jaspionjader/Kosmos-EVAA-immersive-sof-v44-8B | 41e3ecda-8988-456c-b413-19770e2f05c7 | 0.0.1 | hfopenllm_v2/jaspionjader_Kosmos-EVAA-immersive-sof-v44-8B/1762652580.2404292 | 1762652580.24043 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/Kosmos-EVAA-immersive-sof-v44-8B | jaspionjader/Kosmos-EVAA-immersive-sof-v44-8B | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.44078821970150317}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/fr-1-8b | 2014c198-5e12-41ef-8f65-7321d0423573 | 0.0.1 | hfopenllm_v2/jaspionjader_fr-1-8b/1762652580.267912 | 1762652580.2679129 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/fr-1-8b | jaspionjader/fr-1-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.421079402651631}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH"... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/slu-14 | 59703023-61e1-4df0-8542-703d5a318756 | 0.0.1 | hfopenllm_v2/jaspionjader_slu-14/1762652580.27245 | 1762652580.2724512 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/slu-14 | jaspionjader/slu-14 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4106880853912065}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/kstc-4-8b | 6b63598f-4891-4b71-99ca-bc56b780d829 | 0.0.1 | hfopenllm_v2/jaspionjader_kstc-4-8b/1762652580.270735 | 1762652580.270736 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/kstc-4-8b | jaspionjader/kstc-4-8b | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.4769832932175517}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BBH... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 8.03} |
HF Open LLM v2 | jaspionjader | jaspionjader/test-11 | 98f97092-7c95-46dd-94c7-4030f153d197 | 0.0.1 | hfopenllm_v2/jaspionjader_test-11/1762652580.2779882 | 1762652580.2779891 | ["https://open-llm-leaderboard-open-llm-leaderboard.hf.space/api/leaderboard/formatted"] | HF Open LLM v2 | leaderboard | Hugging Face | null | null | third_party | jaspionjader/test-11 | jaspionjader/test-11 | jaspionjader | unknown | [{"evaluation_name": "IFEval", "metric_config": {"evaluation_description": "Accuracy on IFEval", "lower_is_better": false, "score_type": "continuous", "min_score": 0, "max_score": 1}, "score_details": {"score": 0.45412727119598223}}, {"evaluation_name": "BBH", "metric_config": {"evaluation_description": "Accuracy on BB... | {"precision": "bfloat16", "architecture": "LlamaForCausalLM", "params_billions": 4.015} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.