Shawon16 commited on
Commit
1d2faf6
·
verified ·
1 Parent(s): ae02c2f

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ confusion_matrix_kfold_fold_1.tiff filter=lfs diff=lfs merge=lfs -text
37
+ confusion_matrix_test_fold_1.tiff filter=lfs diff=lfs merge=lfs -text
38
+ confusion_matrix_trainfold_fold_1.tiff filter=lfs diff=lfs merge=lfs -text
39
+ long_tail_performance_vs_class_frequency_kfold_fold_1.tiff filter=lfs diff=lfs merge=lfs -text
40
+ long_tail_performance_vs_class_frequency_test_fold_1.tiff filter=lfs diff=lfs merge=lfs -text
41
+ long_tail_performance_vs_class_frequency_trainfold_fold_1.tiff filter=lfs diff=lfs merge=lfs -text
42
+ per_class_accuracy_kfold_fold_1.tiff filter=lfs diff=lfs merge=lfs -text
43
+ per_class_accuracy_test_fold_1.tiff filter=lfs diff=lfs merge=lfs -text
44
+ per_class_accuracy_trainfold_fold_1.tiff filter=lfs diff=lfs merge=lfs -text
confusion_matrix_kfold_fold_1.tiff ADDED

Git LFS Details

  • SHA256: eb44a96f0681d943eb0b9108ddde809429236a65b3ca23115ae0f5d94a585356
  • Pointer size: 133 Bytes
  • Size of remote file: 13 MB
confusion_matrix_test_fold_1.tiff ADDED

Git LFS Details

  • SHA256: 48f9a51c5782edbddee96b1d92e77848a16daa8b797f8ba9e8fb98f9425af214
  • Pointer size: 133 Bytes
  • Size of remote file: 12.9 MB
confusion_matrix_trainfold_fold_1.tiff ADDED

Git LFS Details

  • SHA256: c67420539d72bcb674eea14723e722e692beb72eadd6ff9ed66b5f55f92fef3e
  • Pointer size: 133 Bytes
  • Size of remote file: 13 MB
long_tail_performance_vs_class_frequency_kfold_fold_1.tiff ADDED

Git LFS Details

  • SHA256: 6797f232cbfe5d6e0b17e308bfa079c38e5a170a87604f44ff31b2279502b872
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
long_tail_performance_vs_class_frequency_test_fold_1.tiff ADDED

Git LFS Details

  • SHA256: a22e0503a10c02192a1f2f2b97c01ac2704362115e8dfbd18bade56c1657c612
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
long_tail_performance_vs_class_frequency_trainfold_fold_1.tiff ADDED

Git LFS Details

  • SHA256: e103984712e5b8400ce65d205cddea0460e570d898add2a2f6a049de5ceab0d7
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_kfold_fold_1.tiff ADDED

Git LFS Details

  • SHA256: 515bee8f0dae82e3ec16d850c873f0b5b7e633bd61f0bec760e4bf54dd527a6d
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_test_fold_1.tiff ADDED

Git LFS Details

  • SHA256: e9f93462e61a9e4bbab8bb72962024572fcdc7f6cfbbd8d1314cd19ed63fcb79
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_trainfold_fold_1.tiff ADDED

Git LFS Details

  • SHA256: 5c934b12ed073832dd69c6facb66608e0b8ebf1d2ba2576eec56fdb99f8899ae
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_results_fold_1.csv ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ fold,class,train_accuracy,train_precision,train_recall,train_f1_score,train_top5_accuracy,val_accuracy,val_precision,val_recall,val_f1_score,val_top5_accuracy,test_accuracy,test_precision,test_recall,test_f1_score,test_top5_accuracy
2
+ 1,cake,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9523809523809523,1.0,0.9523809523809523,0.975609756097561,0.9890282131661442
3
+ 1,aids,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.72,1.0,0.8372093023255814,0.9890282131661442
4
+ 1,chikissha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.95,1.0,0.95,0.9743589743589743,0.9890282131661442
5
+ 1,maa,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9523809523809523,1.0,0.975609756097561,0.9890282131661442
6
+ 1,durbol,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
7
+ 1,toothpaste,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
8
+ 1,apartment,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
9
+ 1,chal,1.0,1.0,1.0,1.0,1.0,1.0,0.9230769230769231,1.0,0.96,1.0,0.75,0.9473684210526315,0.75,0.8372093023255814,0.9890282131661442
10
+ 1,chadar,1.0,1.0,1.0,1.0,1.0,0.9333333333333333,1.0,0.9333333333333333,0.9655172413793104,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
11
+ 1,bon,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.4,1.0,0.4,0.5714285714285714,0.9890282131661442
12
+ 1,ac,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9565217391304348,1.0,0.9777777777777777,0.9890282131661442
13
+ 1,konna,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9090909090909091,1.0,0.9523809523809523,0.9890282131661442
14
+ 1,baba,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
15
+ 1,cream,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.7619047619047619,1.0,0.7619047619047619,0.8648648648648649,0.9890282131661442
16
+ 1,chips,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
17
+ 1,chacha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
18
+ 1,dulavai,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
19
+ 1,balu,1.0,0.9841269841269841,1.0,0.992,1.0,1.0,1.0,1.0,1.0,1.0,0.9545454545454546,0.7777777777777778,0.9545454545454546,0.8571428571428571,0.9890282131661442
20
+ 1,baat,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.7727272727272727,1.0,0.7727272727272727,0.8717948717948718,0.9890282131661442
21
+ 1,attio,1.0,0.9904761904761905,1.0,0.9952153110047847,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
22
+ 1,capsule,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
23
+ 1,denadar,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.95,1.0,0.9743589743589743,0.9890282131661442
24
+ 1,biscuts,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
25
+ 1,churi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9090909090909091,1.0,0.9523809523809523,0.9890282131661442
26
+ 1,tattha,0.9913793103448276,1.0,0.9913793103448276,0.9956709956709957,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.8148148148148148,1.0,0.8979591836734694,0.9890282131661442
27
+ 1,clip,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9545454545454546,1.0,0.9545454545454546,0.9767441860465116,0.9890282131661442
28
+ 1,chiruni,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9090909090909091,1.0,0.9523809523809523,0.9890282131661442
29
+ 1,doctor,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
30
+ 1,tv,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.7619047619047619,1.0,0.7619047619047619,0.8648648648648649,0.9890282131661442
31
+ 1,baandej,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
32
+ 1,aam,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.75,1.0,0.8571428571428571,0.9890282131661442
33
+ 1,ayna,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.6818181818181818,0.9375,0.6818181818181818,0.7894736842105263,0.9890282131661442
34
+ 1,alu,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9523809523809523,0.9523809523809523,0.9523809523809523,0.9523809523809523,0.9890282131661442
35
+ 1,daeitto,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
36
+ 1,cha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.4090909090909091,1.0,0.4090909090909091,0.5806451612903226,0.9890282131661442
37
+ 1,audio cassette,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
38
+ 1,dongson,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
39
+ 1,balti,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9523809523809523,1.0,0.975609756097561,0.9890282131661442
40
+ 1,angur,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.6666666666666666,1.0,0.8,0.9890282131661442
41
+ 1,dadi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
42
+ 1,chachi,0.9851851851851852,1.0,0.9851851851851852,0.9925373134328358,1.0,1.0,1.0,1.0,1.0,1.0,0.7272727272727273,0.7272727272727273,0.7272727272727273,0.7272727272727273,0.9890282131661442
43
+ 1,bhai,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
44
+ 1,chosma,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.6086956521739131,1.0,0.6086956521739131,0.7567567567567568,0.9890282131661442
45
+ 1,debor,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
46
+ 1,dal,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9545454545454546,1.0,0.9545454545454546,0.9767441860465116,0.9890282131661442
47
+ 1,bou,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9130434782608695,1.0,0.9545454545454546,0.9890282131661442
48
+ 1,dada,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
49
+ 1,anaros,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.7666666666666667,1.0,0.8679245283018868,0.9890282131661442
50
+ 1,boroi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.84375,1.0,0.9152542372881356,0.9890282131661442
51
+ 1,chokh utha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
52
+ 1,aaple,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
53
+ 1,chocolate,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9166666666666666,1.0,0.9565217391304348,0.9890282131661442
54
+ 1,tshirt,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
55
+ 1,bottam,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
56
+ 1,juta,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.95,0.95,0.95,0.95,0.9890282131661442
57
+ 1,tubelight,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
58
+ 1,tupi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
59
+ 1,dengue,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
60
+ 1,jomoj,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9583333333333334,0.7931034482758621,0.9583333333333334,0.8679245283018868,0.9890282131661442
61
+ 1,chini,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9890282131661442
trainer_state.json ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9987562189054726,
3
+ "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/BdSLW60_10_Fold/VideoMAE_Kinetics_fold__0__BdSLW60_SKF/checkpoint-6324",
4
+ "epoch": 9.099501661129569,
5
+ "eval_steps": 500,
6
+ "global_step": 9030,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.1,
13
+ "grad_norm": 29.418611526489258,
14
+ "learning_rate": 4.9889258028792915e-05,
15
+ "loss": 8.3388,
16
+ "step": 903
17
+ },
18
+ {
19
+ "epoch": 0.1,
20
+ "eval_accuracy": 0.9539800995024875,
21
+ "eval_loss": 0.287722647190094,
22
+ "eval_runtime": 380.8747,
23
+ "eval_samples_per_second": 2.111,
24
+ "eval_steps_per_second": 1.055,
25
+ "step": 903
26
+ },
27
+ {
28
+ "epoch": 1.1000553709856036,
29
+ "grad_norm": 0.4510260224342346,
30
+ "learning_rate": 4.44567491079119e-05,
31
+ "loss": 0.398,
32
+ "step": 1807
33
+ },
34
+ {
35
+ "epoch": 1.1000553709856036,
36
+ "eval_accuracy": 0.9888059701492538,
37
+ "eval_loss": 0.04584308713674545,
38
+ "eval_runtime": 366.4135,
39
+ "eval_samples_per_second": 2.194,
40
+ "eval_steps_per_second": 1.097,
41
+ "step": 1807
42
+ },
43
+ {
44
+ "epoch": 2.1,
45
+ "grad_norm": 0.03591794893145561,
46
+ "learning_rate": 3.890734588409007e-05,
47
+ "loss": 0.0898,
48
+ "step": 2710
49
+ },
50
+ {
51
+ "epoch": 2.1,
52
+ "eval_accuracy": 0.9912935323383084,
53
+ "eval_loss": 0.03670336678624153,
54
+ "eval_runtime": 394.4792,
55
+ "eval_samples_per_second": 2.038,
56
+ "eval_steps_per_second": 1.019,
57
+ "step": 2710
58
+ },
59
+ {
60
+ "epoch": 3.1000553709856034,
61
+ "grad_norm": 0.08393514901399612,
62
+ "learning_rate": 3.334563799680079e-05,
63
+ "loss": 0.0614,
64
+ "step": 3614
65
+ },
66
+ {
67
+ "epoch": 3.1000553709856034,
68
+ "eval_accuracy": 0.9912935323383084,
69
+ "eval_loss": 0.03837832063436508,
70
+ "eval_runtime": 357.3383,
71
+ "eval_samples_per_second": 2.25,
72
+ "eval_steps_per_second": 1.125,
73
+ "step": 3614
74
+ },
75
+ {
76
+ "epoch": 4.1,
77
+ "grad_norm": 0.012894371524453163,
78
+ "learning_rate": 2.7796234772978964e-05,
79
+ "loss": 0.0326,
80
+ "step": 4517
81
+ },
82
+ {
83
+ "epoch": 4.1,
84
+ "eval_accuracy": 0.9950248756218906,
85
+ "eval_loss": 0.02047196589410305,
86
+ "eval_runtime": 384.1902,
87
+ "eval_samples_per_second": 2.093,
88
+ "eval_steps_per_second": 1.046,
89
+ "step": 4517
90
+ },
91
+ {
92
+ "epoch": 5.100055370985603,
93
+ "grad_norm": 0.011511938646435738,
94
+ "learning_rate": 2.2234526885689677e-05,
95
+ "loss": 0.0408,
96
+ "step": 5421
97
+ },
98
+ {
99
+ "epoch": 5.100055370985603,
100
+ "eval_accuracy": 0.9937810945273632,
101
+ "eval_loss": 0.01791091077029705,
102
+ "eval_runtime": 381.8359,
103
+ "eval_samples_per_second": 2.106,
104
+ "eval_steps_per_second": 1.053,
105
+ "step": 5421
106
+ },
107
+ {
108
+ "epoch": 6.1,
109
+ "grad_norm": 0.0027884775772690773,
110
+ "learning_rate": 1.667897133013412e-05,
111
+ "loss": 0.0224,
112
+ "step": 6324
113
+ },
114
+ {
115
+ "epoch": 6.1,
116
+ "eval_accuracy": 0.9987562189054726,
117
+ "eval_loss": 0.00474861916154623,
118
+ "eval_runtime": 201.6524,
119
+ "eval_samples_per_second": 3.987,
120
+ "eval_steps_per_second": 1.994,
121
+ "step": 6324
122
+ },
123
+ {
124
+ "epoch": 7.100055370985603,
125
+ "grad_norm": 0.0023301970213651657,
126
+ "learning_rate": 1.1117263442844839e-05,
127
+ "loss": 0.015,
128
+ "step": 7228
129
+ },
130
+ {
131
+ "epoch": 7.100055370985603,
132
+ "eval_accuracy": 0.9975124378109452,
133
+ "eval_loss": 0.007461219094693661,
134
+ "eval_runtime": 208.3972,
135
+ "eval_samples_per_second": 3.858,
136
+ "eval_steps_per_second": 1.929,
137
+ "step": 7228
138
+ },
139
+ {
140
+ "epoch": 8.1,
141
+ "grad_norm": 0.0014854685869067907,
142
+ "learning_rate": 5.561707887289283e-06,
143
+ "loss": 0.0088,
144
+ "step": 8131
145
+ },
146
+ {
147
+ "epoch": 8.1,
148
+ "eval_accuracy": 0.9975124378109452,
149
+ "eval_loss": 0.009577390737831593,
150
+ "eval_runtime": 203.6802,
151
+ "eval_samples_per_second": 3.947,
152
+ "eval_steps_per_second": 1.974,
153
+ "step": 8131
154
+ },
155
+ {
156
+ "epoch": 9.099501661129569,
157
+ "grad_norm": 0.01574859954416752,
158
+ "learning_rate": 3.076165866863542e-08,
159
+ "loss": 0.0012,
160
+ "step": 9030
161
+ },
162
+ {
163
+ "epoch": 9.099501661129569,
164
+ "eval_accuracy": 0.9987562189054726,
165
+ "eval_loss": 0.004918778780847788,
166
+ "eval_runtime": 209.2235,
167
+ "eval_samples_per_second": 3.843,
168
+ "eval_steps_per_second": 1.921,
169
+ "step": 9030
170
+ },
171
+ {
172
+ "epoch": 9.099501661129569,
173
+ "step": 9030,
174
+ "total_flos": 9.005104637867262e+19,
175
+ "train_loss": 0.9009381401710468,
176
+ "train_runtime": 33313.7237,
177
+ "train_samples_per_second": 2.168,
178
+ "train_steps_per_second": 0.271
179
+ }
180
+ ],
181
+ "logging_steps": 500,
182
+ "max_steps": 9030,
183
+ "num_input_tokens_seen": 0,
184
+ "num_train_epochs": 9223372036854775807,
185
+ "save_steps": 500,
186
+ "stateful_callbacks": {
187
+ "EarlyStoppingCallback": {
188
+ "args": {
189
+ "early_stopping_patience": 5,
190
+ "early_stopping_threshold": 0.0
191
+ },
192
+ "attributes": {
193
+ "early_stopping_patience_counter": 3
194
+ }
195
+ },
196
+ "TrainerControl": {
197
+ "args": {
198
+ "should_epoch_stop": false,
199
+ "should_evaluate": false,
200
+ "should_log": false,
201
+ "should_save": true,
202
+ "should_training_stop": true
203
+ },
204
+ "attributes": {}
205
+ }
206
+ },
207
+ "total_flos": 9.005104637867262e+19,
208
+ "train_batch_size": 2,
209
+ "trial_name": null,
210
+ "trial_params": null
211
+ }