asartipi13 commited on
Commit
dd08229
·
1 Parent(s): c002856

add experiments

Browse files
bl_score_en_fa.csv CHANGED
@@ -1,2 +1,2 @@
1
  ,blue_score
2
- 0,0.19381634891033173
 
1
  ,blue_score
2
+ 0,0.18455731868743896
each_bl_score.csv CHANGED
@@ -1,8 +1,11 @@
1
  ,blue_score
2
- 0,0.08735977858304977
3
- 1,0.10494770854711533
4
- 2,0.11880872398614883
5
- 3,0.1431581676006317
6
- 4,0.17003457248210907
7
- 5,0.1824965476989746
8
- 6,0.19196997582912445
 
 
 
 
1
  ,blue_score
2
+ 0,0.08880972862243652
3
+ 1,0.09895547479391098
4
+ 2,0.11305214464664459
5
+ 3,0.13459543883800507
6
+ 4,0.14872214198112488
7
+ 5,0.1580750197172165
8
+ 6,0.167562797665596
9
+ 7,0.18075799942016602
10
+ 8,0.18740299344062805
11
+ 9,0.1909770667552948
history.csv CHANGED
@@ -1,8 +1,11 @@
1
  ,train_loss,val_loss,time
2
- 0,1.8279560368801806,5.0027263425290585,846.5998046398163
3
- 1,0.8854164853575965,0.9953864651421706,850.0213534832001
4
- 2,0.5886777056808894,0.8349563653270403,848.6701231002808
5
- 3,0.3413949710171555,0.7113787923008204,837.3957936763763
6
- 4,0.1612583426587416,0.6585275561859211,829.6277801990509
7
- 5,0.057262459127873824,0.6881537122341493,850.2859485149384
8
- 6,0.01552234435185628,0.7283899340157708,830.1893310546875
 
 
 
 
1
  ,train_loss,val_loss,time
2
+ 0,1.9561143317697947,5.230697425082326,939.4388282299042
3
+ 1,0.9126862564933634,1.0402087693413098,932.82133436203
4
+ 2,0.6715765313567402,0.8838518025974432,933.0624377727509
5
+ 3,0.39108464309359636,0.770169108795623,931.3709762096405
6
+ 4,0.2479751377555069,0.7448939336463809,926.1014461517334
7
+ 5,0.1459642629989211,0.7201313724120458,929.5383110046387
8
+ 6,0.09117503850909409,0.7622729302383959,802.3399548530579
9
+ 7,0.05941567492331857,0.7234124786530932,797.3547296524048
10
+ 8,0.022946883993646902,0.8209267063066363,795.6876714229584
11
+ 9,0.007395964377629353,0.9265663744881749,793.2133340835571
predicted_en_fa.csv CHANGED
The diff for this file is too large to render. See raw diff
 
total_dev_losss.csv CHANGED
@@ -1,43 +1,61 @@
1
  ,total_dev_losss
2
- 0,23.39219558238983
3
- 1,1.6212423592805862
4
- 2,1.5270865857601166
5
- 3,1.2960425466299057
6
- 4,1.0920665189623833
7
- 5,1.0877244621515274
8
- 6,1.1093337088823318
9
- 7,1.1220472119748592
10
- 8,1.0323783792555332
11
- 9,0.9956429749727249
12
- 10,0.8888876549899578
13
- 11,0.8240288607776165
14
- 12,0.9139084406197071
15
- 13,0.891702588647604
16
- 14,0.8547016531229019
17
- 15,0.8259095698595047
18
- 16,0.7835197560489178
19
- 17,0.7399961836636066
20
- 18,0.7961596436798573
21
- 19,0.7045699674636126
22
- 20,0.7557717487215996
23
- 21,0.696721887215972
24
- 22,0.7490737996995449
25
- 23,0.5659757070243359
26
- 24,0.6970298904925585
27
- 25,0.7088554259389639
28
- 26,0.6630710083991289
29
- 27,0.6744004841893911
30
- 28,0.5621624607592821
31
- 29,0.6456460673362017
32
- 30,0.5411185817793012
33
- 31,0.7370346877723932
34
- 32,0.7417733613401651
35
- 33,0.7273968830704689
36
- 34,0.7538388110697269
37
- 35,0.6277599483728409
38
- 36,0.6866918876767159
39
- 37,0.7586614731699228
40
- 38,0.6794016361236572
41
- 39,0.8491732757538557
42
- 40,0.7519961912184954
43
- 41,0.6444151401519775
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ,total_dev_losss
2
+ 0,23.967587113380432
3
+ 1,1.8037166893482208
4
+ 2,1.4358714148402214
5
+ 3,1.7553121075034142
6
+ 4,1.299879051744938
7
+ 5,1.1218181736767292
8
+ 6,1.1878255866467953
9
+ 7,1.0863665901124477
10
+ 8,1.076308947056532
11
+ 9,1.0020763128995895
12
+ 10,0.9790596328675747
13
+ 11,0.90961554646492
14
+ 12,0.9391063265502453
15
+ 13,1.0305380672216415
16
+ 14,0.9637907557189465
17
+ 15,0.831116322427988
18
+ 16,0.818918738514185
19
+ 17,0.7196406051516533
20
+ 18,0.76445547118783
21
+ 19,0.830062422901392
22
+ 20,0.8638218156993389
23
+ 21,0.7648497447371483
24
+ 22,0.7171620782464743
25
+ 23,0.6806631200015545
26
+ 24,0.7112197764217854
27
+ 25,0.7234151568263769
28
+ 26,0.8256546352058649
29
+ 27,0.8299507386982441
30
+ 28,0.7095781825482845
31
+ 29,0.6695451121777296
32
+ 30,0.6676106099039316
33
+ 31,0.8043257333338261
34
+ 32,0.664452601224184
35
+ 33,0.8083262089639902
36
+ 34,0.7853506226092577
37
+ 35,0.5907224584370852
38
+ 36,0.7328919470310211
39
+ 37,0.7920571155846119
40
+ 38,0.6919057555496693
41
+ 39,0.8423732127994299
42
+ 40,0.7910541761666536
43
+ 41,0.7233553742989898
44
+ 42,0.6646301914006472
45
+ 43,0.6859897579997778
46
+ 44,0.8358445167541504
47
+ 45,0.7685340624302626
48
+ 46,0.717962983995676
49
+ 47,0.6675133593380451
50
+ 48,0.7882163543254137
51
+ 49,0.8395467530936003
52
+ 50,0.7894728928804398
53
+ 51,0.8263336010277271
54
+ 52,0.8610436785966158
55
+ 53,0.8209469579160213
56
+ 54,0.8475340958684683
57
+ 55,0.8468272816389799
58
+ 56,1.0151942446827888
59
+ 57,0.8960121870040894
60
+ 58,1.0333637073636055
61
+ 59,0.9204667303711176
total_train_losss.csv CHANGED
The diff for this file is too large to render. See raw diff