Upload folder using huggingface_hub (#460)
Browse files- 6644414eb41856aa01671574e02794671bdf235a6e7f0895f64bdc17710ed9ed (76433988883a3dd981b8a9cc97e1e639c378f220)
Pretrain_language_model/save_final/slimpajama_moe_no_attmoe_660M_standardlb/tmp/result-model-400000.pth.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"val/loss": 2.252077859545511,
|
| 3 |
+
"val/accuracy": 0.5291467091393849,
|
| 4 |
+
"val/perplexity": 9.507470514675877,
|
| 5 |
+
"val/time_since_best_loss": 0,
|
| 6 |
+
"val/time_since_best_accuracy": 0,
|
| 7 |
+
"mean_accuracy": 0.5291467091393849,
|
| 8 |
+
"mean_loss": 2.252077859545511,
|
| 9 |
+
"boolq/accuracy/dev": 0.6214067278287462,
|
| 10 |
+
"boolq/accuracy/group_average": 0.6214067278287462,
|
| 11 |
+
"boolq/accuracy/seq_average": 0.6214067278287462
|
| 12 |
+
}
|