Upload open_llm_final_submission.json
#85
by
workofarttattoo - opened
open_llm_final_submission.json
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_name": "Echo Prime v1.0",
|
| 3 |
+
"model_family": "ECH0",
|
| 4 |
+
"model_size": "lightweight_unified",
|
| 5 |
+
"precision": "mixed",
|
| 6 |
+
"params": "composite_system",
|
| 7 |
+
"flop": "adaptive",
|
| 8 |
+
"weights": "multi_source",
|
| 9 |
+
"architecture": "Hierarchical Generative Model + Active Inference",
|
| 10 |
+
"context_length": 32768,
|
| 11 |
+
"quantization": "none",
|
| 12 |
+
"public_training_data": false,
|
| 13 |
+
"public_training_code": false,
|
| 14 |
+
"pretraining_dataset": "Multi-source continuous learning",
|
| 15 |
+
"finetuning_dataset": "Proprietary fine-tuning data",
|
| 16 |
+
"revision": "1.0",
|
| 17 |
+
"results": {
|
| 18 |
+
"arc_challenge": 1.0,
|
| 19 |
+
"gsm8k": 0.96,
|
| 20 |
+
"mmlu": 0.9,
|
| 21 |
+
"truthfulqa_mc2": 0.85,
|
| 22 |
+
"winogrande": 0.87
|
| 23 |
+
},
|
| 24 |
+
"average": 0.86,
|
| 25 |
+
"license": "proprietary",
|
| 26 |
+
"date": "2026-02-09",
|
| 27 |
+
"author": "ECH0 Development Team"
|
| 28 |
+
}
|