asartipi13 commited on
Commit
89cc3e8
·
1 Parent(s): 4aa980f

add experiments

Browse files
bl_score_fa_en.csv CHANGED
@@ -1,2 +1,2 @@
1
  ,blue_score
2
- 0,0.16135020554065704
 
1
  ,blue_score
2
+ 0,0.1592547446489334
each_bl_score.csv CHANGED
@@ -1,8 +1,11 @@
1
  ,blue_score
2
- 0,0.11859533190727234
3
- 1,0.13332048058509827
4
- 2,0.14318153262138367
5
- 3,0.1495945006608963
6
- 4,0.15948884189128876
7
- 5,0.1656816452741623
8
- 6,0.1654045134782791
 
 
 
 
1
  ,blue_score
2
+ 0,0.11670329421758652
3
+ 1,0.12396610528230667
4
+ 2,0.13726304471492767
5
+ 3,0.14831164479255676
6
+ 4,0.1515844613313675
7
+ 5,0.15619590878486633
8
+ 6,0.15876658260822296
9
+ 7,0.1609804332256317
10
+ 8,0.1624710112810135
11
+ 9,0.16187584400177002
history.csv CHANGED
@@ -1,8 +1,11 @@
1
  ,train_loss,val_loss,time
2
- 0,1.4625656938545426,4.872440102199714,763.9421217441559
3
- 1,0.746155915647773,0.8548438331733147,758.5161001682281
4
- 2,0.553410593753442,0.7596966611842314,742.1053085327148
5
- 3,0.38919247734227186,0.7348400220895807,755.4473373889923
6
- 4,0.24015845642092815,0.7410638000195225,758.8557472229004
7
- 5,0.13064575631574293,0.8180758859962225,762.4018065929413
8
- 6,0.06076858585642929,0.9606465436518192,762.1522171497345
 
 
 
 
1
  ,train_loss,val_loss,time
2
+ 0,1.6915148982470651,4.832429361219208,800.5257730484009
3
+ 1,0.7864536325508189,0.9086142542461554,794.7516753673553
4
+ 2,0.5718621290732767,0.8024971981843313,789.880786895752
5
+ 3,0.4082096520056977,0.7835246470446388,926.4268462657928
6
+ 4,0.2961512892324234,0.7893346262474855,926.0841906070709
7
+ 5,0.2209102366850755,0.8380610874543587,926.664252281189
8
+ 6,0.149750809134724,0.86142563726753,926.6233110427856
9
+ 7,0.10069031567309876,0.9578769095242023,926.9494485855103
10
+ 8,0.06267556542780169,1.1554853779574235,927.7087087631226
11
+ 9,0.0356921887437224,1.2907919051746528,929.8957777023315
predicted_fa_en.csv CHANGED
The diff for this file is too large to render. See raw diff
 
total_dev_losss.csv CHANGED
@@ -1,43 +1,61 @@
1
  ,total_dev_losss
2
- 0,23.57292139530182
3
- 1,1.4040356278419495
4
- 2,1.1943675056099892
5
- 3,1.063797164708376
6
- 4,1.0258639976382256
7
- 5,0.9736549220979214
8
- 6,0.8770406022667885
9
- 7,0.8575083464384079
10
- 8,0.9290241785347462
11
- 9,0.854883223772049
12
- 10,0.8328331932425499
13
- 11,0.777773454785347
14
- 12,0.8052382841706276
15
- 13,0.799306109547615
16
- 14,0.7689665965735912
17
- 15,0.7369842305779457
18
- 16,0.7257743291556835
19
- 17,0.7219104170799255
20
- 18,0.7551550529897213
21
- 19,0.7404764071106911
22
- 20,0.799946878105402
23
- 21,0.7616524305194616
24
- 22,0.6583556439727545
25
- 23,0.6934537198394537
26
- 24,0.6937616653740406
27
- 25,0.8076251931488514
28
- 26,0.7571748867630959
29
- 27,0.8200835958123207
30
- 28,0.6796118784695864
31
- 29,0.6881255805492401
32
- 30,0.7088275365531445
33
- 31,0.9157954789698124
34
- 32,0.7888336293399334
35
- 33,0.7995687685906887
36
- 34,0.8343023806810379
37
- 35,0.8611275218427181
38
- 36,0.762531653046608
39
- 37,0.9590636417269707
40
- 38,0.974993109703064
41
- 39,0.9177718795835972
42
- 40,1.0606913268566132
43
- 41,1.0888276509940624
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ,total_dev_losss
2
+ 0,23.181299448013306
3
+ 1,1.4837843626737595
4
+ 2,1.2423470504581928
5
+ 3,1.065871998667717
6
+ 4,1.0497450716793537
7
+ 5,0.971528235822916
8
+ 6,1.0016193576157093
9
+ 7,0.8927176035940647
10
+ 8,0.9750287644565105
11
+ 9,0.875748798251152
12
+ 10,0.8883978873491287
13
+ 11,0.8181731142103672
14
+ 12,0.8870008550584316
15
+ 13,0.8385061137378216
16
+ 14,0.85753083974123
17
+ 15,0.7312553264200687
18
+ 16,0.7775383703410625
19
+ 17,0.723151683807373
20
+ 18,0.8400003686547279
21
+ 19,0.8064890280365944
22
+ 20,0.839827436953783
23
+ 21,0.7770166806876659
24
+ 22,0.756560355424881
25
+ 23,0.6812540125101805
26
+ 24,0.7287239395081997
27
+ 25,0.8207096569240093
28
+ 26,0.788226805627346
29
+ 27,0.77645668014884
30
+ 28,0.8333783559501171
31
+ 29,0.7885123193264008
32
+ 30,0.8056388981640339
33
+ 31,0.8312023989856243
34
+ 32,0.8840327486395836
35
+ 33,0.8848282136023045
36
+ 34,0.8308812193572521
37
+ 35,0.791783045977354
38
+ 36,0.7729857675731182
39
+ 37,0.8665478080511093
40
+ 38,0.8901644647121429
41
+ 39,0.892485748976469
42
+ 40,0.889591783285141
43
+ 41,0.8567782510071993
44
+ 42,0.9051547795534134
45
+ 43,1.0061692520976067
46
+ 44,0.9185377284884453
47
+ 45,0.9955393373966217
48
+ 46,1.012190781533718
49
+ 47,0.9096695780754089
50
+ 48,0.9458565674722195
51
+ 49,1.2680066749453545
52
+ 50,1.1704692170023918
53
+ 51,1.2183393016457558
54
+ 52,1.1790268868207932
55
+ 53,1.1512136198580265
56
+ 54,1.19849618896842
57
+ 55,1.4313330426812172
58
+ 56,1.1859880648553371
59
+ 57,1.3486342057585716
60
+ 58,1.3298128433525562
61
+ 59,1.2504870854318142
total_train_losss.csv CHANGED
The diff for this file is too large to render. See raw diff