Upload checkpoints/train_logs.txt with huggingface_hub
Browse files- checkpoints/train_logs.txt +20 -80
checkpoints/train_logs.txt
CHANGED
|
@@ -1,80 +1,20 @@
|
|
| 1 |
-
epoch:0 | train loss:
|
| 2 |
-
epoch:1 | train loss: 0.
|
| 3 |
-
epoch:2 | train loss: 0.
|
| 4 |
-
epoch:3 | train loss: 0.
|
| 5 |
-
epoch:4 | train loss: 0.
|
| 6 |
-
epoch:5 | train loss: 0.
|
| 7 |
-
epoch:6 | train loss: 0.
|
| 8 |
-
epoch:7 | train loss: 0.
|
| 9 |
-
epoch:8 | train loss: 0.
|
| 10 |
-
epoch:9 | train loss: 0.
|
| 11 |
-
epoch:10 | train loss: 0.
|
| 12 |
-
epoch:11 | train loss: 0.
|
| 13 |
-
epoch:12 | train loss: 0.
|
| 14 |
-
epoch:13 | train loss: 0.
|
| 15 |
-
epoch:14 | train loss: 0.
|
| 16 |
-
epoch:15 | train loss: 0.
|
| 17 |
-
epoch:16 | train loss: 0.
|
| 18 |
-
epoch:17 | train loss: 0.
|
| 19 |
-
epoch:18 | train loss: 0.
|
| 20 |
-
epoch:19 | train loss: 0.
|
| 21 |
-
epoch:20 | train loss: 0.343086 | val_loss:0.339160
|
| 22 |
-
epoch:21 | train loss: 0.353715 | val_loss:0.313373
|
| 23 |
-
epoch:22 | train loss: 0.346464 | val_loss:0.349604
|
| 24 |
-
epoch:23 | train loss: 0.346148 | val_loss:0.338928
|
| 25 |
-
epoch:24 | train loss: 0.343711 | val_loss:0.358593
|
| 26 |
-
epoch:25 | train loss: 0.344212 | val_loss:0.344940
|
| 27 |
-
epoch:26 | train loss: 0.336860 | val_loss:0.323936
|
| 28 |
-
epoch:27 | train loss: 0.342427 | val_loss:0.350062
|
| 29 |
-
epoch:28 | train loss: 0.342525 | val_loss:0.318503
|
| 30 |
-
epoch:29 | train loss: 0.343273 | val_loss:0.353699
|
| 31 |
-
epoch:30 | train loss: 0.335665 | val_loss:0.351245
|
| 32 |
-
epoch:31 | train loss: 0.333117 | val_loss:0.337402
|
| 33 |
-
epoch:32 | train loss: 0.341981 | val_loss:0.338283
|
| 34 |
-
epoch:33 | train loss: 0.332584 | val_loss:0.322904
|
| 35 |
-
epoch:34 | train loss: 0.344967 | val_loss:0.337000
|
| 36 |
-
epoch:35 | train loss: 0.335911 | val_loss:0.339282
|
| 37 |
-
epoch:36 | train loss: 0.336757 | val_loss:0.346806
|
| 38 |
-
epoch:37 | train loss: 0.335283 | val_loss:0.336629
|
| 39 |
-
epoch:38 | train loss: 0.341270 | val_loss:0.347422
|
| 40 |
-
epoch:39 | train loss: 0.331153 | val_loss:0.317502
|
| 41 |
-
epoch:40 | train loss: 0.335883 | val_loss:0.327056
|
| 42 |
-
epoch:41 | train loss: 0.336514 | val_loss:0.327075
|
| 43 |
-
epoch:42 | train loss: 0.333111 | val_loss:0.335423
|
| 44 |
-
epoch:43 | train loss: 0.329750 | val_loss:0.323630
|
| 45 |
-
epoch:44 | train loss: 0.328020 | val_loss:0.341802
|
| 46 |
-
epoch:45 | train loss: 0.336057 | val_loss:0.352105
|
| 47 |
-
epoch:46 | train loss: 0.333871 | val_loss:0.332672
|
| 48 |
-
epoch:47 | train loss: 0.328025 | val_loss:0.320695
|
| 49 |
-
epoch:48 | train loss: 0.328654 | val_loss:0.331762
|
| 50 |
-
epoch:49 | train loss: 0.333991 | val_loss:0.300728
|
| 51 |
-
epoch:50 | train loss: 0.328299 | val_loss:0.341610
|
| 52 |
-
epoch:51 | train loss: 0.326566 | val_loss:0.328439
|
| 53 |
-
epoch:52 | train loss: 0.323551 | val_loss:0.331395
|
| 54 |
-
epoch:53 | train loss: 0.326660 | val_loss:0.357509
|
| 55 |
-
epoch:54 | train loss: 0.326364 | val_loss:0.348237
|
| 56 |
-
epoch:55 | train loss: 0.325146 | val_loss:0.352877
|
| 57 |
-
epoch:56 | train loss: 0.326764 | val_loss:0.319848
|
| 58 |
-
epoch:57 | train loss: 0.330026 | val_loss:0.330971
|
| 59 |
-
epoch:58 | train loss: 0.324813 | val_loss:0.313570
|
| 60 |
-
epoch:59 | train loss: 0.326677 | val_loss:0.291725
|
| 61 |
-
epoch:60 | train loss: 0.331478 | val_loss:0.337928
|
| 62 |
-
epoch:61 | train loss: 0.318888 | val_loss:0.340632
|
| 63 |
-
epoch:62 | train loss: 0.323466 | val_loss:0.329508
|
| 64 |
-
epoch:63 | train loss: 0.327050 | val_loss:0.317438
|
| 65 |
-
epoch:64 | train loss: 0.331010 | val_loss:0.322188
|
| 66 |
-
epoch:65 | train loss: 0.319511 | val_loss:0.323175
|
| 67 |
-
epoch:66 | train loss: 0.327105 | val_loss:0.353646
|
| 68 |
-
epoch:67 | train loss: 0.318625 | val_loss:0.317440
|
| 69 |
-
epoch:68 | train loss: 0.324964 | val_loss:0.334219
|
| 70 |
-
epoch:69 | train loss: 0.325614 | val_loss:0.337909
|
| 71 |
-
epoch:70 | train loss: 0.320022 | val_loss:0.317496
|
| 72 |
-
epoch:71 | train loss: 0.324041 | val_loss:0.285073
|
| 73 |
-
epoch:72 | train loss: 0.323589 | val_loss:0.312426
|
| 74 |
-
epoch:73 | train loss: 0.314499 | val_loss:0.339005
|
| 75 |
-
epoch:74 | train loss: 0.320577 | val_loss:0.315502
|
| 76 |
-
epoch:75 | train loss: 0.322160 | val_loss:0.328063
|
| 77 |
-
epoch:76 | train loss: 0.320567 | val_loss:0.332127
|
| 78 |
-
epoch:77 | train loss: 0.323989 | val_loss:0.325119
|
| 79 |
-
epoch:78 | train loss: 0.317423 | val_loss:0.343041
|
| 80 |
-
epoch:79 | train loss: 0.324545 | val_loss:0.314887
|
|
|
|
| 1 |
+
epoch:0 | train loss: 1.073054 | val_loss:0.448881
|
| 2 |
+
epoch:1 | train loss: 0.392142 | val_loss:0.399242
|
| 3 |
+
epoch:2 | train loss: 0.371812 | val_loss:0.393474
|
| 4 |
+
epoch:3 | train loss: 0.364254 | val_loss:0.362560
|
| 5 |
+
epoch:4 | train loss: 0.352752 | val_loss:0.360939
|
| 6 |
+
epoch:5 | train loss: 0.344565 | val_loss:0.354360
|
| 7 |
+
epoch:6 | train loss: 0.350875 | val_loss:0.361243
|
| 8 |
+
epoch:7 | train loss: 0.352274 | val_loss:0.371514
|
| 9 |
+
epoch:8 | train loss: 0.344127 | val_loss:0.358672
|
| 10 |
+
epoch:9 | train loss: 0.349167 | val_loss:0.367300
|
| 11 |
+
epoch:10 | train loss: 0.352776 | val_loss:0.352947
|
| 12 |
+
epoch:11 | train loss: 0.335472 | val_loss:0.331375
|
| 13 |
+
epoch:12 | train loss: 0.343942 | val_loss:0.334605
|
| 14 |
+
epoch:13 | train loss: 0.343223 | val_loss:0.339721
|
| 15 |
+
epoch:14 | train loss: 0.349869 | val_loss:0.343825
|
| 16 |
+
epoch:15 | train loss: 0.340668 | val_loss:0.368064
|
| 17 |
+
epoch:16 | train loss: 0.340238 | val_loss:0.333418
|
| 18 |
+
epoch:17 | train loss: 0.337068 | val_loss:0.351729
|
| 19 |
+
epoch:18 | train loss: 0.335290 | val_loss:0.344091
|
| 20 |
+
epoch:19 | train loss: 0.334789 | val_loss:0.341503
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|