Shawon16 commited on
Commit
95cfb7c
·
verified ·
1 Parent(s): 8b581a1

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ confusion_matrix_kfold_fold_2.tiff filter=lfs diff=lfs merge=lfs -text
37
+ confusion_matrix_test_fold_2.tiff filter=lfs diff=lfs merge=lfs -text
38
+ confusion_matrix_trainfold_fold_2.tiff filter=lfs diff=lfs merge=lfs -text
39
+ long_tail_performance_vs_class_frequency_kfold_fold_2.tiff filter=lfs diff=lfs merge=lfs -text
40
+ long_tail_performance_vs_class_frequency_test_fold_2.tiff filter=lfs diff=lfs merge=lfs -text
41
+ long_tail_performance_vs_class_frequency_trainfold_fold_2.tiff filter=lfs diff=lfs merge=lfs -text
42
+ per_class_accuracy_kfold_fold_2.tiff filter=lfs diff=lfs merge=lfs -text
43
+ per_class_accuracy_test_fold_2.tiff filter=lfs diff=lfs merge=lfs -text
44
+ per_class_accuracy_trainfold_fold_2.tiff filter=lfs diff=lfs merge=lfs -text
confusion_matrix_kfold_fold_2.tiff ADDED

Git LFS Details

  • SHA256: 95736f01ca2692271e11654ec68a83f02c50aa1addbd534911659f227b404db3
  • Pointer size: 133 Bytes
  • Size of remote file: 13 MB
confusion_matrix_test_fold_2.tiff ADDED

Git LFS Details

  • SHA256: 790d364ce82e629da231abc11e464acc410fae3a798784637f1f3b94d43fec8b
  • Pointer size: 133 Bytes
  • Size of remote file: 12.9 MB
confusion_matrix_trainfold_fold_2.tiff ADDED

Git LFS Details

  • SHA256: 48ee081b8f6ac6f93d8a5f736c9f74c7f635f6b478ffb8c8c46dd5ac29c16832
  • Pointer size: 133 Bytes
  • Size of remote file: 13 MB
long_tail_performance_vs_class_frequency_kfold_fold_2.tiff ADDED

Git LFS Details

  • SHA256: 85050f581557e78a0368e502cb5e1cdaf33a694a1554e501635e6255d9da2458
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
long_tail_performance_vs_class_frequency_test_fold_2.tiff ADDED

Git LFS Details

  • SHA256: 79a04e9de54d476eba0eea0649dc9e85621db28b613cd056c9c99e84cf404330
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
long_tail_performance_vs_class_frequency_trainfold_fold_2.tiff ADDED

Git LFS Details

  • SHA256: ae74056b5f9ac78989955db6d289b245695e0b7b84c72133c97f830c4e1557fd
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_kfold_fold_2.tiff ADDED

Git LFS Details

  • SHA256: a2718688ee15a8d27b711744182d4b8f6de8c78850dcb013e1db627d68af0217
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_test_fold_2.tiff ADDED

Git LFS Details

  • SHA256: 80e2a7509373e479400041524f0186f27a0f643ced07d15c5797516dfd225b01
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_trainfold_fold_2.tiff ADDED

Git LFS Details

  • SHA256: f7384efd152b886d4c2c5c9a8d4d50562a238d36b524016e2059665348958792
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_results_fold_2.csv ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ fold,class,train_accuracy,train_precision,train_recall,train_f1_score,train_top5_accuracy,val_accuracy,val_precision,val_recall,val_f1_score,val_top5_accuracy,test_accuracy,test_precision,test_recall,test_f1_score,test_top5_accuracy
2
+ 2,cake,0.9920634920634921,1.0,0.9920634920634921,0.9960159362549801,1.0,1.0,1.0,1.0,1.0,1.0,0.9523809523809523,1.0,0.9523809523809523,0.975609756097561,0.9913793103448276
3
+ 2,aids,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.5555555555555556,0.7142857142857143,0.5555555555555556,0.625,0.9913793103448276
4
+ 2,chikissha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.85,1.0,0.85,0.918918918918919,0.9913793103448276
5
+ 2,maa,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9090909090909091,1.0,0.9523809523809523,0.9913793103448276
6
+ 2,durbol,1.0,0.9920634920634921,1.0,0.9960159362549801,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9565217391304348,1.0,0.9777777777777777,0.9913793103448276
7
+ 2,toothpaste,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.92,1.0,0.9583333333333334,0.9913793103448276
8
+ 2,apartment,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.95,1.0,0.9743589743589743,0.9913793103448276
9
+ 2,chal,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.7916666666666666,0.95,0.7916666666666666,0.8636363636363636,0.9913793103448276
10
+ 2,chadar,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.6818181818181818,1.0,0.6818181818181818,0.8108108108108109,0.9913793103448276
11
+ 2,bon,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.75,0.8333333333333334,0.75,0.7894736842105263,0.9913793103448276
12
+ 2,ac,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.7333333333333333,1.0,0.8461538461538461,0.9913793103448276
13
+ 2,konna,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
14
+ 2,baba,0.9938271604938271,1.0,0.9938271604938271,0.9969040247678018,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
15
+ 2,cream,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.47619047619047616,1.0,0.47619047619047616,0.6451612903225806,0.9913793103448276
16
+ 2,chips,1.0,1.0,1.0,1.0,1.0,1.0,0.9166666666666666,1.0,0.9565217391304348,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
17
+ 2,chacha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
18
+ 2,dulavai,1.0,1.0,1.0,1.0,1.0,0.9285714285714286,1.0,0.9285714285714286,0.9629629629629629,1.0,1.0,0.8518518518518519,1.0,0.92,0.9913793103448276
19
+ 2,balu,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9545454545454546,0.9545454545454546,0.9545454545454546,0.9545454545454546,0.9913793103448276
20
+ 2,baat,1.0,1.0,1.0,1.0,1.0,1.0,0.9333333333333333,1.0,0.9655172413793104,1.0,0.7727272727272727,1.0,0.7727272727272727,0.8717948717948718,0.9913793103448276
21
+ 2,attio,1.0,0.9811320754716981,1.0,0.9904761904761905,1.0,1.0,1.0,1.0,1.0,1.0,0.7727272727272727,0.9444444444444444,0.7727272727272727,0.85,0.9913793103448276
22
+ 2,capsule,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
23
+ 2,denadar,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.95,1.0,0.9743589743589743,0.9913793103448276
24
+ 2,biscuts,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9523809523809523,1.0,0.975609756097561,0.9913793103448276
25
+ 2,churi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9090909090909091,1.0,0.9523809523809523,0.9913793103448276
26
+ 2,tattha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.8461538461538461,1.0,0.9166666666666666,0.9913793103448276
27
+ 2,clip,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
28
+ 2,chiruni,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9523809523809523,1.0,0.975609756097561,0.9913793103448276
29
+ 2,doctor,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
30
+ 2,tv,0.9919354838709677,1.0,0.9919354838709677,0.9959514170040485,1.0,1.0,1.0,1.0,1.0,1.0,0.9047619047619048,1.0,0.9047619047619048,0.95,0.9913793103448276
31
+ 2,baandej,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
32
+ 2,aam,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.84,1.0,0.9130434782608695,0.9913793103448276
33
+ 2,ayna,0.991869918699187,1.0,0.991869918699187,0.9959183673469387,1.0,1.0,1.0,1.0,1.0,1.0,0.6818181818181818,1.0,0.6818181818181818,0.8108108108108109,0.9913793103448276
34
+ 2,alu,0.989010989010989,0.989010989010989,0.989010989010989,0.989010989010989,1.0,1.0,0.9166666666666666,1.0,0.9565217391304348,1.0,1.0,0.9545454545454546,1.0,0.9767441860465116,0.9913793103448276
35
+ 2,daeitto,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
36
+ 2,cha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.5,1.0,0.5,0.6666666666666666,0.9913793103448276
37
+ 2,audio cassette,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
38
+ 2,dongson,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
39
+ 2,balti,1.0,0.9893617021276596,1.0,0.9946524064171123,1.0,0.9,1.0,0.9,0.9473684210526315,1.0,0.95,1.0,0.95,0.9743589743589743,0.9913793103448276
40
+ 2,angur,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.72,1.0,0.8372093023255814,0.9913793103448276
41
+ 2,dadi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9583333333333334,1.0,0.9787234042553191,0.9913793103448276
42
+ 2,chachi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.6666666666666666,1.0,0.8,0.9913793103448276
43
+ 2,bhai,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
44
+ 2,chosma,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9565217391304348,0.88,0.9565217391304348,0.9166666666666666,0.9913793103448276
45
+ 2,debor,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
46
+ 2,dal,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9090909090909091,0.9523809523809523,0.9090909090909091,0.9302325581395349,0.9913793103448276
47
+ 2,bou,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9130434782608695,1.0,0.9545454545454546,0.9913793103448276
48
+ 2,dada,1.0,0.9937106918238994,1.0,0.9968454258675079,1.0,1.0,1.0,1.0,1.0,1.0,0.9047619047619048,0.7307692307692307,0.9047619047619048,0.8085106382978723,0.9913793103448276
49
+ 2,anaros,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9565217391304348,1.0,0.9565217391304348,0.9777777777777777,0.9913793103448276
50
+ 2,boroi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9629629629629629,0.8387096774193549,0.9629629629629629,0.896551724137931,0.9913793103448276
51
+ 2,chokh utha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
52
+ 2,aaple,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
53
+ 2,chocolate,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9545454545454546,0.9130434782608695,0.9545454545454546,0.9333333333333333,0.9913793103448276
54
+ 2,tshirt,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
55
+ 2,bottam,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.95,1.0,0.95,0.9743589743589743,0.9913793103448276
56
+ 2,juta,1.0,1.0,1.0,1.0,1.0,0.9090909090909091,1.0,0.9090909090909091,0.9523809523809523,1.0,0.9,0.9,0.9,0.9,0.9913793103448276
57
+ 2,tubelight,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9545454545454546,1.0,0.9767441860465116,0.9913793103448276
58
+ 2,tupi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
59
+ 2,dengue,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
60
+ 2,jomoj,0.9900990099009901,1.0,0.9900990099009901,0.9950248756218906,1.0,1.0,1.0,1.0,1.0,1.0,0.9583333333333334,1.0,0.9583333333333334,0.9787234042553191,0.9913793103448276
61
+ 2,chini,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9913793103448276
trainer_state.json ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9962640099626401,
3
+ "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/BdSLW60_10_Fold/VideoMAE_Kinetics_fold__1__BdSLW60_SKF/checkpoint-5421",
4
+ "epoch": 9.099501661129569,
5
+ "eval_steps": 500,
6
+ "global_step": 9030,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.1,
13
+ "grad_norm": 20.25380516052246,
14
+ "learning_rate": 4.9889258028792915e-05,
15
+ "loss": 8.3842,
16
+ "step": 903
17
+ },
18
+ {
19
+ "epoch": 0.1,
20
+ "eval_accuracy": 0.9601494396014943,
21
+ "eval_loss": 0.261745423078537,
22
+ "eval_runtime": 175.8411,
23
+ "eval_samples_per_second": 4.567,
24
+ "eval_steps_per_second": 2.286,
25
+ "step": 903
26
+ },
27
+ {
28
+ "epoch": 1.1000553709856036,
29
+ "grad_norm": 9.491084098815918,
30
+ "learning_rate": 4.44567491079119e-05,
31
+ "loss": 0.3809,
32
+ "step": 1807
33
+ },
34
+ {
35
+ "epoch": 1.1000553709856036,
36
+ "eval_accuracy": 0.9937733499377335,
37
+ "eval_loss": 0.02950221486389637,
38
+ "eval_runtime": 171.8238,
39
+ "eval_samples_per_second": 4.673,
40
+ "eval_steps_per_second": 2.34,
41
+ "step": 1807
42
+ },
43
+ {
44
+ "epoch": 2.1,
45
+ "grad_norm": 0.05116604268550873,
46
+ "learning_rate": 3.8901193552356345e-05,
47
+ "loss": 0.0785,
48
+ "step": 2710
49
+ },
50
+ {
51
+ "epoch": 2.1,
52
+ "eval_accuracy": 0.9863013698630136,
53
+ "eval_loss": 0.057530105113983154,
54
+ "eval_runtime": 175.8249,
55
+ "eval_samples_per_second": 4.567,
56
+ "eval_steps_per_second": 2.286,
57
+ "step": 2710
58
+ },
59
+ {
60
+ "epoch": 3.1000553709856034,
61
+ "grad_norm": 0.02004808746278286,
62
+ "learning_rate": 3.334563799680079e-05,
63
+ "loss": 0.0559,
64
+ "step": 3614
65
+ },
66
+ {
67
+ "epoch": 3.1000553709856034,
68
+ "eval_accuracy": 0.9887920298879203,
69
+ "eval_loss": 0.044266898185014725,
70
+ "eval_runtime": 168.2715,
71
+ "eval_samples_per_second": 4.772,
72
+ "eval_steps_per_second": 2.389,
73
+ "step": 3614
74
+ },
75
+ {
76
+ "epoch": 4.1,
77
+ "grad_norm": 0.034268010407686234,
78
+ "learning_rate": 2.7796234772978964e-05,
79
+ "loss": 0.0343,
80
+ "step": 4517
81
+ },
82
+ {
83
+ "epoch": 4.1,
84
+ "eval_accuracy": 0.9838107098381071,
85
+ "eval_loss": 0.07113216072320938,
86
+ "eval_runtime": 170.6766,
87
+ "eval_samples_per_second": 4.705,
88
+ "eval_steps_per_second": 2.355,
89
+ "step": 4517
90
+ },
91
+ {
92
+ "epoch": 5.100055370985603,
93
+ "grad_norm": 0.006388051901012659,
94
+ "learning_rate": 2.2234526885689677e-05,
95
+ "loss": 0.0242,
96
+ "step": 5421
97
+ },
98
+ {
99
+ "epoch": 5.100055370985603,
100
+ "eval_accuracy": 0.9962640099626401,
101
+ "eval_loss": 0.021267583593726158,
102
+ "eval_runtime": 165.7809,
103
+ "eval_samples_per_second": 4.844,
104
+ "eval_steps_per_second": 2.425,
105
+ "step": 5421
106
+ },
107
+ {
108
+ "epoch": 6.1,
109
+ "grad_norm": 0.007680265232920647,
110
+ "learning_rate": 1.667897133013412e-05,
111
+ "loss": 0.0217,
112
+ "step": 6324
113
+ },
114
+ {
115
+ "epoch": 6.1,
116
+ "eval_accuracy": 0.9937733499377335,
117
+ "eval_loss": 0.030281249433755875,
118
+ "eval_runtime": 171.3026,
119
+ "eval_samples_per_second": 4.688,
120
+ "eval_steps_per_second": 2.347,
121
+ "step": 6324
122
+ },
123
+ {
124
+ "epoch": 7.100055370985603,
125
+ "grad_norm": 0.0014248612569645047,
126
+ "learning_rate": 1.1117263442844839e-05,
127
+ "loss": 0.0056,
128
+ "step": 7228
129
+ },
130
+ {
131
+ "epoch": 7.100055370985603,
132
+ "eval_accuracy": 0.9962640099626401,
133
+ "eval_loss": 0.021625792607665062,
134
+ "eval_runtime": 164.3393,
135
+ "eval_samples_per_second": 4.886,
136
+ "eval_steps_per_second": 2.446,
137
+ "step": 7228
138
+ },
139
+ {
140
+ "epoch": 8.1,
141
+ "grad_norm": 0.049291741102933884,
142
+ "learning_rate": 5.561707887289283e-06,
143
+ "loss": 0.0067,
144
+ "step": 8131
145
+ },
146
+ {
147
+ "epoch": 8.1,
148
+ "eval_accuracy": 0.9962640099626401,
149
+ "eval_loss": 0.027468109503388405,
150
+ "eval_runtime": 169.3911,
151
+ "eval_samples_per_second": 4.741,
152
+ "eval_steps_per_second": 2.373,
153
+ "step": 8131
154
+ },
155
+ {
156
+ "epoch": 9.099501661129569,
157
+ "grad_norm": 0.0021701958030462265,
158
+ "learning_rate": 3.076165866863542e-08,
159
+ "loss": 0.001,
160
+ "step": 9030
161
+ },
162
+ {
163
+ "epoch": 9.099501661129569,
164
+ "eval_accuracy": 0.9962640099626401,
165
+ "eval_loss": 0.017694512382149696,
166
+ "eval_runtime": 174.2614,
167
+ "eval_samples_per_second": 4.608,
168
+ "eval_steps_per_second": 2.307,
169
+ "step": 9030
170
+ },
171
+ {
172
+ "epoch": 9.099501661129569,
173
+ "step": 9030,
174
+ "total_flos": 9.006226676074414e+19,
175
+ "train_loss": 0.8993382768773557,
176
+ "train_runtime": 20553.9951,
177
+ "train_samples_per_second": 3.515,
178
+ "train_steps_per_second": 0.439
179
+ }
180
+ ],
181
+ "logging_steps": 500,
182
+ "max_steps": 9030,
183
+ "num_input_tokens_seen": 0,
184
+ "num_train_epochs": 9223372036854775807,
185
+ "save_steps": 500,
186
+ "stateful_callbacks": {
187
+ "EarlyStoppingCallback": {
188
+ "args": {
189
+ "early_stopping_patience": 5,
190
+ "early_stopping_threshold": 0.0
191
+ },
192
+ "attributes": {
193
+ "early_stopping_patience_counter": 4
194
+ }
195
+ },
196
+ "TrainerControl": {
197
+ "args": {
198
+ "should_epoch_stop": false,
199
+ "should_evaluate": false,
200
+ "should_log": false,
201
+ "should_save": true,
202
+ "should_training_stop": true
203
+ },
204
+ "attributes": {}
205
+ }
206
+ },
207
+ "total_flos": 9.006226676074414e+19,
208
+ "train_batch_size": 2,
209
+ "trial_name": null,
210
+ "trial_params": null
211
+ }