Shawon16 commited on
Commit
fb13514
·
verified ·
1 Parent(s): 0c3b928

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ confusion_matrix_kfold_fold_7.tiff filter=lfs diff=lfs merge=lfs -text
37
+ confusion_matrix_test_fold_7.tiff filter=lfs diff=lfs merge=lfs -text
38
+ confusion_matrix_trainfold_fold_7.tiff filter=lfs diff=lfs merge=lfs -text
39
+ long_tail_performance_vs_class_frequency_kfold_fold_7.tiff filter=lfs diff=lfs merge=lfs -text
40
+ long_tail_performance_vs_class_frequency_test_fold_7.tiff filter=lfs diff=lfs merge=lfs -text
41
+ long_tail_performance_vs_class_frequency_trainfold_fold_7.tiff filter=lfs diff=lfs merge=lfs -text
42
+ per_class_accuracy_kfold_fold_7.tiff filter=lfs diff=lfs merge=lfs -text
43
+ per_class_accuracy_test_fold_7.tiff filter=lfs diff=lfs merge=lfs -text
44
+ per_class_accuracy_trainfold_fold_7.tiff filter=lfs diff=lfs merge=lfs -text
confusion_matrix_kfold_fold_7.tiff ADDED

Git LFS Details

  • SHA256: 72cebd9d3f9da5f6314845e2e06bee8f045889c64402d73d1f4e72854f5a161a
  • Pointer size: 133 Bytes
  • Size of remote file: 13 MB
confusion_matrix_test_fold_7.tiff ADDED

Git LFS Details

  • SHA256: 0bcc1492d74bb5d8fe92c3f457139faa774b0445669c3dbc69bcba84d3a3f648
  • Pointer size: 133 Bytes
  • Size of remote file: 12.9 MB
confusion_matrix_trainfold_fold_7.tiff ADDED

Git LFS Details

  • SHA256: 006f190fffbe378e96b4f5050f2d759d895ef2df70aa0a0f66109a101822224f
  • Pointer size: 133 Bytes
  • Size of remote file: 13 MB
long_tail_performance_vs_class_frequency_kfold_fold_7.tiff ADDED

Git LFS Details

  • SHA256: 0b8cdfba39909c48260cc5165d03ab425b93d0733aa4485b58ca7a39d8f121af
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
long_tail_performance_vs_class_frequency_test_fold_7.tiff ADDED

Git LFS Details

  • SHA256: 26fe2ff7259153d851f36e89af938bd466dd7d56d59df99526f5e42b6998870f
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
long_tail_performance_vs_class_frequency_trainfold_fold_7.tiff ADDED

Git LFS Details

  • SHA256: a88af18b274dfc8bc101e0320bf8b0554e7c3ac4f7fc6e9df49ac57fd1b31b21
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_kfold_fold_7.tiff ADDED

Git LFS Details

  • SHA256: 5c9c8bc9932c632fa93bc6f00cb394dfebd6a2decc7706ffc3b0e48ead751ade
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_test_fold_7.tiff ADDED

Git LFS Details

  • SHA256: 853f35fd850d575a20f2b418375f75f31a7aa06d6c67617e1bd7cce24ea7e5cd
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_trainfold_fold_7.tiff ADDED

Git LFS Details

  • SHA256: 12944a596daa771b40c4fec8ce7dbb668c6f9815b2bed1132d5813e3ba67d47d
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_results_fold_7.csv ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ fold,class,train_accuracy,train_precision,train_recall,train_f1_score,train_top5_accuracy,val_accuracy,val_precision,val_recall,val_f1_score,val_top5_accuracy,test_accuracy,test_precision,test_recall,test_f1_score,test_top5_accuracy
2
+ 7,cake,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9523809523809523,1.0,0.9523809523809523,0.975609756097561,0.9913793103448276
3
+ 7,aids,1.0,0.9924812030075187,1.0,0.9962264150943396,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.782608695652174,1.0,0.8780487804878049,0.9913793103448276
4
+ 7,chikissha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.85,1.0,0.85,0.918918918918919,0.9913793103448276
5
+ 7,maa,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9090909090909091,1.0,0.9523809523809523,0.9913793103448276
6
+ 7,durbol,0.992,1.0,0.992,0.9959839357429718,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.8636363636363636,0.95,0.8636363636363636,0.9047619047619048,0.9913793103448276
7
+ 7,toothpaste,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9583333333333334,1.0,0.9787234042553191,0.9913793103448276
8
+ 7,apartment,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
9
+ 7,chal,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.7916666666666666,1.0,0.7916666666666666,0.8837209302325582,0.9913793103448276
10
+ 7,chadar,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9565217391304348,1.0,0.9777777777777777,0.9913793103448276
11
+ 7,bon,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.75,1.0,0.75,0.8571428571428571,0.9913793103448276
12
+ 7,ac,1.0,0.9910714285714286,1.0,0.9955156950672646,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
13
+ 7,konna,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9523809523809523,1.0,0.975609756097561,0.9913793103448276
14
+ 7,baba,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9583333333333334,1.0,0.9787234042553191,0.9913793103448276
15
+ 7,cream,1.0,1.0,1.0,1.0,1.0,0.8571428571428571,1.0,0.8571428571428571,0.9230769230769231,0.9975093399750934,0.7619047619047619,1.0,0.7619047619047619,0.8648648648648649,0.9913793103448276
16
+ 7,chips,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
17
+ 7,chacha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
18
+ 7,dulavai,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.7931034482758621,1.0,0.8846153846153846,0.9913793103448276
19
+ 7,balu,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9545454545454546,1.0,0.9545454545454546,0.9767441860465116,0.9913793103448276
20
+ 7,baat,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
21
+ 7,attio,1.0,1.0,1.0,1.0,1.0,1.0,0.9230769230769231,1.0,0.96,0.9975093399750934,0.8636363636363636,0.8636363636363636,0.8636363636363636,0.8636363636363636,0.9913793103448276
22
+ 7,capsule,1.0,1.0,1.0,1.0,1.0,1.0,0.9166666666666666,1.0,0.9565217391304348,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
23
+ 7,denadar,1.0,1.0,1.0,1.0,1.0,0.9285714285714286,1.0,0.9285714285714286,0.9629629629629629,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
24
+ 7,biscuts,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
25
+ 7,churi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
26
+ 7,tattha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
27
+ 7,clip,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.8148148148148148,1.0,0.8979591836734694,0.9913793103448276
28
+ 7,chiruni,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9523809523809523,1.0,0.975609756097561,0.9913793103448276
29
+ 7,doctor,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
30
+ 7,tv,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9523809523809523,1.0,0.9523809523809523,0.975609756097561,0.9913793103448276
31
+ 7,baandej,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.95,1.0,0.9743589743589743,0.9913793103448276
32
+ 7,aam,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
33
+ 7,ayna,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
34
+ 7,alu,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.875,1.0,0.9333333333333333,0.9913793103448276
35
+ 7,daeitto,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
36
+ 7,cha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.6363636363636364,0.9333333333333333,0.6363636363636364,0.7567567567567568,0.9913793103448276
37
+ 7,audio cassette,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
38
+ 7,dongson,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
39
+ 7,balti,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.85,0.9444444444444444,0.85,0.8947368421052632,0.9913793103448276
40
+ 7,angur,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
41
+ 7,dadi,0.9927536231884058,1.0,0.9927536231884058,0.9963636363636363,1.0,0.9375,0.9375,0.9375,0.9375,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
42
+ 7,chachi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.7272727272727273,0.8,0.7272727272727273,0.7619047619047619,0.9913793103448276
43
+ 7,bhai,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
44
+ 7,chosma,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.8695652173913043,0.9523809523809523,0.8695652173913043,0.9090909090909091,0.9913793103448276
45
+ 7,debor,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.92,0.9583333333333334,0.92,0.9387755102040817,0.9913793103448276
46
+ 7,dal,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9545454545454546,1.0,0.9545454545454546,0.9767441860465116,0.9913793103448276
47
+ 7,bou,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.7241379310344828,1.0,0.84,0.9913793103448276
48
+ 7,dada,1.0,1.0,1.0,1.0,1.0,1.0,0.9473684210526315,1.0,0.972972972972973,0.9975093399750934,1.0,0.9545454545454546,1.0,0.9767441860465116,0.9913793103448276
49
+ 7,anaros,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9130434782608695,0.875,0.9130434782608695,0.8936170212765957,0.9913793103448276
50
+ 7,boroi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9259259259259259,0.9259259259259259,0.9259259259259259,0.9259259259259259,0.9913793103448276
51
+ 7,chokh utha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
52
+ 7,aaple,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
53
+ 7,chocolate,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9565217391304348,1.0,0.9777777777777777,0.9913793103448276
54
+ 7,tshirt,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
55
+ 7,bottam,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.95,1.0,0.95,0.9743589743589743,0.9913793103448276
56
+ 7,juta,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.95,0.95,0.95,0.95,0.9913793103448276
57
+ 7,tubelight,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
58
+ 7,tupi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
59
+ 7,dengue,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
60
+ 7,jomoj,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9583333333333334,0.92,0.9583333333333334,0.9387755102040817,0.9913793103448276
61
+ 7,chini,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9913793103448276
trainer_state.json ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9950186799501868,
3
+ "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/BdSLW60_10_Fold/VideoMAE_Kinetics_fold__6__BdSLW60_SKF/checkpoint-6324",
4
+ "epoch": 9.099501661129569,
5
+ "eval_steps": 500,
6
+ "global_step": 9030,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.1,
13
+ "grad_norm": 40.80965805053711,
14
+ "learning_rate": 4.9889258028792915e-05,
15
+ "loss": 8.274,
16
+ "step": 903
17
+ },
18
+ {
19
+ "epoch": 0.1,
20
+ "eval_accuracy": 0.9676214196762142,
21
+ "eval_loss": 0.271345317363739,
22
+ "eval_runtime": 206.0084,
23
+ "eval_samples_per_second": 3.898,
24
+ "eval_steps_per_second": 1.951,
25
+ "step": 903
26
+ },
27
+ {
28
+ "epoch": 1.1000553709856036,
29
+ "grad_norm": 0.12732265889644623,
30
+ "learning_rate": 4.44567491079119e-05,
31
+ "loss": 0.3704,
32
+ "step": 1807
33
+ },
34
+ {
35
+ "epoch": 1.1000553709856036,
36
+ "eval_accuracy": 0.9925280199252802,
37
+ "eval_loss": 0.036753587424755096,
38
+ "eval_runtime": 166.3615,
39
+ "eval_samples_per_second": 4.827,
40
+ "eval_steps_per_second": 2.416,
41
+ "step": 1807
42
+ },
43
+ {
44
+ "epoch": 2.1,
45
+ "grad_norm": 2.391895055770874,
46
+ "learning_rate": 3.890734588409007e-05,
47
+ "loss": 0.0847,
48
+ "step": 2710
49
+ },
50
+ {
51
+ "epoch": 2.1,
52
+ "eval_accuracy": 0.9887920298879203,
53
+ "eval_loss": 0.0403018444776535,
54
+ "eval_runtime": 170.7649,
55
+ "eval_samples_per_second": 4.702,
56
+ "eval_steps_per_second": 2.354,
57
+ "step": 2710
58
+ },
59
+ {
60
+ "epoch": 3.1000553709856034,
61
+ "grad_norm": 0.022390121594071388,
62
+ "learning_rate": 3.334563799680079e-05,
63
+ "loss": 0.0547,
64
+ "step": 3614
65
+ },
66
+ {
67
+ "epoch": 3.1000553709856034,
68
+ "eval_accuracy": 0.9925280199252802,
69
+ "eval_loss": 0.0509362556040287,
70
+ "eval_runtime": 222.2424,
71
+ "eval_samples_per_second": 3.613,
72
+ "eval_steps_per_second": 1.809,
73
+ "step": 3614
74
+ },
75
+ {
76
+ "epoch": 4.1,
77
+ "grad_norm": 0.00862208567559719,
78
+ "learning_rate": 2.7790082441245237e-05,
79
+ "loss": 0.0472,
80
+ "step": 4517
81
+ },
82
+ {
83
+ "epoch": 4.1,
84
+ "eval_accuracy": 0.9863013698630136,
85
+ "eval_loss": 0.0508890263736248,
86
+ "eval_runtime": 168.5494,
87
+ "eval_samples_per_second": 4.764,
88
+ "eval_steps_per_second": 2.385,
89
+ "step": 4517
90
+ },
91
+ {
92
+ "epoch": 5.100055370985603,
93
+ "grad_norm": 0.005311078857630491,
94
+ "learning_rate": 2.222837455395595e-05,
95
+ "loss": 0.026,
96
+ "step": 5421
97
+ },
98
+ {
99
+ "epoch": 5.100055370985603,
100
+ "eval_accuracy": 0.9900373599003736,
101
+ "eval_loss": 0.05522875860333443,
102
+ "eval_runtime": 165.4109,
103
+ "eval_samples_per_second": 4.855,
104
+ "eval_steps_per_second": 2.43,
105
+ "step": 5421
106
+ },
107
+ {
108
+ "epoch": 6.1,
109
+ "grad_norm": 34.34551239013672,
110
+ "learning_rate": 1.6672818998400395e-05,
111
+ "loss": 0.0266,
112
+ "step": 6324
113
+ },
114
+ {
115
+ "epoch": 6.1,
116
+ "eval_accuracy": 0.9950186799501868,
117
+ "eval_loss": 0.02956746704876423,
118
+ "eval_runtime": 168.6884,
119
+ "eval_samples_per_second": 4.76,
120
+ "eval_steps_per_second": 2.383,
121
+ "step": 6324
122
+ },
123
+ {
124
+ "epoch": 7.100055370985603,
125
+ "grad_norm": 0.001888299360871315,
126
+ "learning_rate": 1.1117263442844839e-05,
127
+ "loss": 0.0147,
128
+ "step": 7228
129
+ },
130
+ {
131
+ "epoch": 7.100055370985603,
132
+ "eval_accuracy": 0.9925280199252802,
133
+ "eval_loss": 0.037218641489744186,
134
+ "eval_runtime": 188.0222,
135
+ "eval_samples_per_second": 4.271,
136
+ "eval_steps_per_second": 2.138,
137
+ "step": 7228
138
+ },
139
+ {
140
+ "epoch": 8.1,
141
+ "grad_norm": 0.001524563180282712,
142
+ "learning_rate": 5.561707887289283e-06,
143
+ "loss": 0.0004,
144
+ "step": 8131
145
+ },
146
+ {
147
+ "epoch": 8.1,
148
+ "eval_accuracy": 0.9925280199252802,
149
+ "eval_loss": 0.03954409062862396,
150
+ "eval_runtime": 180.5518,
151
+ "eval_samples_per_second": 4.447,
152
+ "eval_steps_per_second": 2.227,
153
+ "step": 8131
154
+ },
155
+ {
156
+ "epoch": 9.099501661129569,
157
+ "grad_norm": 0.0018760759849101305,
158
+ "learning_rate": 3.076165866863542e-08,
159
+ "loss": 0.0011,
160
+ "step": 9030
161
+ },
162
+ {
163
+ "epoch": 9.099501661129569,
164
+ "eval_accuracy": 0.9900373599003736,
165
+ "eval_loss": 0.04370718449354172,
166
+ "eval_runtime": 164.8466,
167
+ "eval_samples_per_second": 4.871,
168
+ "eval_steps_per_second": 2.439,
169
+ "step": 9030
170
+ },
171
+ {
172
+ "epoch": 9.099501661129569,
173
+ "step": 9030,
174
+ "total_flos": 9.006226676074414e+19,
175
+ "train_loss": 0.8900273501344957,
176
+ "train_runtime": 19437.2006,
177
+ "train_samples_per_second": 3.717,
178
+ "train_steps_per_second": 0.465
179
+ }
180
+ ],
181
+ "logging_steps": 500,
182
+ "max_steps": 9030,
183
+ "num_input_tokens_seen": 0,
184
+ "num_train_epochs": 9223372036854775807,
185
+ "save_steps": 500,
186
+ "stateful_callbacks": {
187
+ "EarlyStoppingCallback": {
188
+ "args": {
189
+ "early_stopping_patience": 5,
190
+ "early_stopping_threshold": 0.0
191
+ },
192
+ "attributes": {
193
+ "early_stopping_patience_counter": 3
194
+ }
195
+ },
196
+ "TrainerControl": {
197
+ "args": {
198
+ "should_epoch_stop": false,
199
+ "should_evaluate": false,
200
+ "should_log": false,
201
+ "should_save": true,
202
+ "should_training_stop": true
203
+ },
204
+ "attributes": {}
205
+ }
206
+ },
207
+ "total_flos": 9.006226676074414e+19,
208
+ "train_batch_size": 2,
209
+ "trial_name": null,
210
+ "trial_params": null
211
+ }