Muhammed164 commited on
Commit
5ff1d9b
·
verified ·
1 Parent(s): de9e99f

Upload checkpoints/train_logs.txt with huggingface_hub

Browse files
Files changed (1) hide show
  1. checkpoints/train_logs.txt +20 -80
checkpoints/train_logs.txt CHANGED
@@ -1,80 +1,20 @@
1
- epoch:0 | train loss: 2.488488 | val_loss:0.673000
2
- epoch:1 | train loss: 0.562669 | val_loss:0.471757
3
- epoch:2 | train loss: 0.426671 | val_loss:0.389693
4
- epoch:3 | train loss: 0.393032 | val_loss:0.387741
5
- epoch:4 | train loss: 0.384743 | val_loss:0.389496
6
- epoch:5 | train loss: 0.383109 | val_loss:0.384204
7
- epoch:6 | train loss: 0.372450 | val_loss:0.401784
8
- epoch:7 | train loss: 0.374903 | val_loss:0.381014
9
- epoch:8 | train loss: 0.369852 | val_loss:0.353790
10
- epoch:9 | train loss: 0.369738 | val_loss:0.346133
11
- epoch:10 | train loss: 0.368340 | val_loss:0.346144
12
- epoch:11 | train loss: 0.368726 | val_loss:0.384480
13
- epoch:12 | train loss: 0.361943 | val_loss:0.354854
14
- epoch:13 | train loss: 0.355516 | val_loss:0.355703
15
- epoch:14 | train loss: 0.355792 | val_loss:0.361314
16
- epoch:15 | train loss: 0.358746 | val_loss:0.348872
17
- epoch:16 | train loss: 0.351417 | val_loss:0.364919
18
- epoch:17 | train loss: 0.344650 | val_loss:0.378364
19
- epoch:18 | train loss: 0.350553 | val_loss:0.378952
20
- epoch:19 | train loss: 0.346499 | val_loss:0.351281
21
- epoch:20 | train loss: 0.343086 | val_loss:0.339160
22
- epoch:21 | train loss: 0.353715 | val_loss:0.313373
23
- epoch:22 | train loss: 0.346464 | val_loss:0.349604
24
- epoch:23 | train loss: 0.346148 | val_loss:0.338928
25
- epoch:24 | train loss: 0.343711 | val_loss:0.358593
26
- epoch:25 | train loss: 0.344212 | val_loss:0.344940
27
- epoch:26 | train loss: 0.336860 | val_loss:0.323936
28
- epoch:27 | train loss: 0.342427 | val_loss:0.350062
29
- epoch:28 | train loss: 0.342525 | val_loss:0.318503
30
- epoch:29 | train loss: 0.343273 | val_loss:0.353699
31
- epoch:30 | train loss: 0.335665 | val_loss:0.351245
32
- epoch:31 | train loss: 0.333117 | val_loss:0.337402
33
- epoch:32 | train loss: 0.341981 | val_loss:0.338283
34
- epoch:33 | train loss: 0.332584 | val_loss:0.322904
35
- epoch:34 | train loss: 0.344967 | val_loss:0.337000
36
- epoch:35 | train loss: 0.335911 | val_loss:0.339282
37
- epoch:36 | train loss: 0.336757 | val_loss:0.346806
38
- epoch:37 | train loss: 0.335283 | val_loss:0.336629
39
- epoch:38 | train loss: 0.341270 | val_loss:0.347422
40
- epoch:39 | train loss: 0.331153 | val_loss:0.317502
41
- epoch:40 | train loss: 0.335883 | val_loss:0.327056
42
- epoch:41 | train loss: 0.336514 | val_loss:0.327075
43
- epoch:42 | train loss: 0.333111 | val_loss:0.335423
44
- epoch:43 | train loss: 0.329750 | val_loss:0.323630
45
- epoch:44 | train loss: 0.328020 | val_loss:0.341802
46
- epoch:45 | train loss: 0.336057 | val_loss:0.352105
47
- epoch:46 | train loss: 0.333871 | val_loss:0.332672
48
- epoch:47 | train loss: 0.328025 | val_loss:0.320695
49
- epoch:48 | train loss: 0.328654 | val_loss:0.331762
50
- epoch:49 | train loss: 0.333991 | val_loss:0.300728
51
- epoch:50 | train loss: 0.328299 | val_loss:0.341610
52
- epoch:51 | train loss: 0.326566 | val_loss:0.328439
53
- epoch:52 | train loss: 0.323551 | val_loss:0.331395
54
- epoch:53 | train loss: 0.326660 | val_loss:0.357509
55
- epoch:54 | train loss: 0.326364 | val_loss:0.348237
56
- epoch:55 | train loss: 0.325146 | val_loss:0.352877
57
- epoch:56 | train loss: 0.326764 | val_loss:0.319848
58
- epoch:57 | train loss: 0.330026 | val_loss:0.330971
59
- epoch:58 | train loss: 0.324813 | val_loss:0.313570
60
- epoch:59 | train loss: 0.326677 | val_loss:0.291725
61
- epoch:60 | train loss: 0.331478 | val_loss:0.337928
62
- epoch:61 | train loss: 0.318888 | val_loss:0.340632
63
- epoch:62 | train loss: 0.323466 | val_loss:0.329508
64
- epoch:63 | train loss: 0.327050 | val_loss:0.317438
65
- epoch:64 | train loss: 0.331010 | val_loss:0.322188
66
- epoch:65 | train loss: 0.319511 | val_loss:0.323175
67
- epoch:66 | train loss: 0.327105 | val_loss:0.353646
68
- epoch:67 | train loss: 0.318625 | val_loss:0.317440
69
- epoch:68 | train loss: 0.324964 | val_loss:0.334219
70
- epoch:69 | train loss: 0.325614 | val_loss:0.337909
71
- epoch:70 | train loss: 0.320022 | val_loss:0.317496
72
- epoch:71 | train loss: 0.324041 | val_loss:0.285073
73
- epoch:72 | train loss: 0.323589 | val_loss:0.312426
74
- epoch:73 | train loss: 0.314499 | val_loss:0.339005
75
- epoch:74 | train loss: 0.320577 | val_loss:0.315502
76
- epoch:75 | train loss: 0.322160 | val_loss:0.328063
77
- epoch:76 | train loss: 0.320567 | val_loss:0.332127
78
- epoch:77 | train loss: 0.323989 | val_loss:0.325119
79
- epoch:78 | train loss: 0.317423 | val_loss:0.343041
80
- epoch:79 | train loss: 0.324545 | val_loss:0.314887
 
1
+ epoch:0 | train loss: 1.073054 | val_loss:0.448881
2
+ epoch:1 | train loss: 0.392142 | val_loss:0.399242
3
+ epoch:2 | train loss: 0.371812 | val_loss:0.393474
4
+ epoch:3 | train loss: 0.364254 | val_loss:0.362560
5
+ epoch:4 | train loss: 0.352752 | val_loss:0.360939
6
+ epoch:5 | train loss: 0.344565 | val_loss:0.354360
7
+ epoch:6 | train loss: 0.350875 | val_loss:0.361243
8
+ epoch:7 | train loss: 0.352274 | val_loss:0.371514
9
+ epoch:8 | train loss: 0.344127 | val_loss:0.358672
10
+ epoch:9 | train loss: 0.349167 | val_loss:0.367300
11
+ epoch:10 | train loss: 0.352776 | val_loss:0.352947
12
+ epoch:11 | train loss: 0.335472 | val_loss:0.331375
13
+ epoch:12 | train loss: 0.343942 | val_loss:0.334605
14
+ epoch:13 | train loss: 0.343223 | val_loss:0.339721
15
+ epoch:14 | train loss: 0.349869 | val_loss:0.343825
16
+ epoch:15 | train loss: 0.340668 | val_loss:0.368064
17
+ epoch:16 | train loss: 0.340238 | val_loss:0.333418
18
+ epoch:17 | train loss: 0.337068 | val_loss:0.351729
19
+ epoch:18 | train loss: 0.335290 | val_loss:0.344091
20
+ epoch:19 | train loss: 0.334789 | val_loss:0.341503