Shawon16 commited on
Commit
1b804a4
·
verified ·
1 Parent(s): e25f8fd

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ confusion_matrix_kfold_fold_6.tiff filter=lfs diff=lfs merge=lfs -text
37
+ confusion_matrix_test_fold_6.tiff filter=lfs diff=lfs merge=lfs -text
38
+ confusion_matrix_trainfold_fold_6.tiff filter=lfs diff=lfs merge=lfs -text
39
+ long_tail_performance_vs_class_frequency_kfold_fold_6.tiff filter=lfs diff=lfs merge=lfs -text
40
+ long_tail_performance_vs_class_frequency_test_fold_6.tiff filter=lfs diff=lfs merge=lfs -text
41
+ long_tail_performance_vs_class_frequency_trainfold_fold_6.tiff filter=lfs diff=lfs merge=lfs -text
42
+ per_class_accuracy_kfold_fold_6.tiff filter=lfs diff=lfs merge=lfs -text
43
+ per_class_accuracy_test_fold_6.tiff filter=lfs diff=lfs merge=lfs -text
44
+ per_class_accuracy_trainfold_fold_6.tiff filter=lfs diff=lfs merge=lfs -text
confusion_matrix_kfold_fold_6.tiff ADDED

Git LFS Details

  • SHA256: 23610acbd52699a0bc43f00cea78c2ce66dc06c07a6cf213047735b56283e60c
  • Pointer size: 133 Bytes
  • Size of remote file: 13 MB
confusion_matrix_test_fold_6.tiff ADDED

Git LFS Details

  • SHA256: 87ef3b3515ad0fe79d1fcf5289d2789dd24e628e19a9281255d3ef661158bc78
  • Pointer size: 133 Bytes
  • Size of remote file: 12.9 MB
confusion_matrix_trainfold_fold_6.tiff ADDED

Git LFS Details

  • SHA256: 6ec4c8ddb48313c4b8e5ff3462ce2b48e80f0a900787b4655a77fdfc1033c61e
  • Pointer size: 133 Bytes
  • Size of remote file: 13 MB
long_tail_performance_vs_class_frequency_kfold_fold_6.tiff ADDED

Git LFS Details

  • SHA256: 2115312659529ebd71fc4868a4de4472cb8fb84e8df352fa5570fd86b6e4db63
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
long_tail_performance_vs_class_frequency_test_fold_6.tiff ADDED

Git LFS Details

  • SHA256: 4acae5009596810ca0daf59c23afa2837b6b336d13a0159bf1e17eba66009e0a
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
long_tail_performance_vs_class_frequency_trainfold_fold_6.tiff ADDED

Git LFS Details

  • SHA256: ff48ed162ce7a311b83a951a921d26673011ae4047207df57d5ca43848c0e28b
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_kfold_fold_6.tiff ADDED

Git LFS Details

  • SHA256: efba6a06818ffaa9bd2b6361368783b04128805eaeb64ad3b24497c3ae8ed91c
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_test_fold_6.tiff ADDED

Git LFS Details

  • SHA256: a78fb56c9e3012541cec19447c5052fbe49bc46e0469ba540b3a0d453fbe8ef9
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_trainfold_fold_6.tiff ADDED

Git LFS Details

  • SHA256: 00bbca8ff09829027d40db6f0ee843bce145dbc03653240fefb17f9c9b236f38
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_results_fold_6.csv ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ fold,class,train_accuracy,train_precision,train_recall,train_f1_score,train_top5_accuracy,val_accuracy,val_precision,val_recall,val_f1_score,val_top5_accuracy,test_accuracy,test_precision,test_recall,test_f1_score,test_top5_accuracy
2
+ 6,cake,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9523809523809523,1.0,0.9523809523809523,0.975609756097561,0.9905956112852664
3
+ 6,aids,1.0,0.9850746268656716,1.0,0.9924812030075187,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.8571428571428571,1.0,0.9230769230769231,0.9905956112852664
4
+ 6,chikissha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.85,1.0,0.85,0.918918918918919,0.9905956112852664
5
+ 6,maa,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.8695652173913043,1.0,0.9302325581395349,0.9905956112852664
6
+ 6,durbol,0.992,1.0,0.992,0.9959839357429718,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
7
+ 6,toothpaste,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
8
+ 6,apartment,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.95,1.0,0.9743589743589743,0.9905956112852664
9
+ 6,chal,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.7083333333333334,0.9444444444444444,0.7083333333333334,0.8095238095238095,0.9905956112852664
10
+ 6,chadar,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
11
+ 6,bon,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9,0.9,0.9,0.9,0.9905956112852664
12
+ 6,ac,1.0,1.0,1.0,1.0,1.0,1.0,0.9230769230769231,1.0,0.96,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
13
+ 6,konna,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9523809523809523,1.0,0.975609756097561,0.9905956112852664
14
+ 6,baba,0.9938271604938271,1.0,0.9938271604938271,0.9969040247678018,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
15
+ 6,cream,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.6190476190476191,1.0,0.6190476190476191,0.7647058823529411,0.9905956112852664
16
+ 6,chips,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
17
+ 6,chacha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
18
+ 6,dulavai,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.92,1.0,0.9583333333333334,0.9905956112852664
19
+ 6,balu,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9545454545454546,0.875,0.9545454545454546,0.9130434782608695,0.9905956112852664
20
+ 6,baat,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.88,1.0,0.9361702127659575,0.9905956112852664
21
+ 6,attio,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9090909090909091,0.9523809523809523,0.9090909090909091,0.9302325581395349,0.9905956112852664
22
+ 6,capsule,1.0,1.0,1.0,1.0,1.0,1.0,0.9166666666666666,1.0,0.9565217391304348,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
23
+ 6,denadar,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
24
+ 6,biscuts,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
25
+ 6,churi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
26
+ 6,tattha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9166666666666666,1.0,0.9565217391304348,0.9905956112852664
27
+ 6,clip,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
28
+ 6,chiruni,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
29
+ 6,doctor,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
30
+ 6,tv,1.0,1.0,1.0,1.0,1.0,0.9285714285714286,1.0,0.9285714285714286,0.9629629629629629,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
31
+ 6,baandej,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
32
+ 6,aam,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
33
+ 6,ayna,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9090909090909091,1.0,0.9090909090909091,0.9523809523809523,0.9905956112852664
34
+ 6,alu,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9523809523809523,0.9523809523809523,0.9523809523809523,0.9523809523809523,0.9905956112852664
35
+ 6,daeitto,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
36
+ 6,cha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
37
+ 6,audio cassette,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
38
+ 6,dongson,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9473684210526315,1.0,0.9473684210526315,0.972972972972973,0.9905956112852664
39
+ 6,balti,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.95,0.95,0.95,0.95,0.9905956112852664
40
+ 6,angur,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
41
+ 6,dadi,0.9927536231884058,1.0,0.9927536231884058,0.9963636363636363,1.0,0.9375,1.0,0.9375,0.967741935483871,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
42
+ 6,chachi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.8461538461538461,1.0,0.9166666666666666,0.9905956112852664
43
+ 6,bhai,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
44
+ 6,chosma,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.8695652173913043,1.0,0.8695652173913043,0.9302325581395349,0.9905956112852664
45
+ 6,debor,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.92,0.9583333333333334,0.92,0.9387755102040817,0.9905956112852664
46
+ 6,dal,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9545454545454546,1.0,0.9545454545454546,0.9767441860465116,0.9905956112852664
47
+ 6,bou,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9545454545454546,1.0,0.9767441860465116,0.9905956112852664
48
+ 6,dada,1.0,0.9936708860759493,1.0,0.9968253968253968,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9545454545454546,1.0,0.9767441860465116,0.9905956112852664
49
+ 6,anaros,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.92,1.0,0.9583333333333334,0.9905956112852664
50
+ 6,boroi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9629629629629629,0.9629629629629629,0.9629629629629629,0.9629629629629629,0.9905956112852664
51
+ 6,chokh utha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
52
+ 6,aaple,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
53
+ 6,chocolate,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.88,1.0,0.9361702127659575,0.9905956112852664
54
+ 6,tshirt,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,0.9545454545454546,1.0,0.9767441860465116,0.9905956112852664
55
+ 6,bottam,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.95,1.0,0.95,0.9743589743589743,0.9905956112852664
56
+ 6,juta,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.95,0.95,0.95,0.95,0.9905956112852664
57
+ 6,tubelight,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
58
+ 6,tupi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
59
+ 6,dengue,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
60
+ 6,jomoj,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,0.9583333333333334,1.0,0.9583333333333334,0.9787234042553191,0.9905956112852664
61
+ 6,chini,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9975093399750934,1.0,1.0,1.0,1.0,0.9905956112852664
trainer_state.json ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9975093399750934,
3
+ "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/BdSLW60_10_Fold/VideoMAE_Kinetics_fold__5__BdSLW60_SKF/checkpoint-8131",
4
+ "epoch": 9.099501661129569,
5
+ "eval_steps": 500,
6
+ "global_step": 9030,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.1,
13
+ "grad_norm": 29.523759841918945,
14
+ "learning_rate": 4.9889258028792915e-05,
15
+ "loss": 8.2749,
16
+ "step": 903
17
+ },
18
+ {
19
+ "epoch": 0.1,
20
+ "eval_accuracy": 0.9638854296388543,
21
+ "eval_loss": 0.25073474645614624,
22
+ "eval_runtime": 216.8383,
23
+ "eval_samples_per_second": 3.703,
24
+ "eval_steps_per_second": 1.854,
25
+ "step": 903
26
+ },
27
+ {
28
+ "epoch": 1.1000553709856036,
29
+ "grad_norm": 0.20937076210975647,
30
+ "learning_rate": 4.44567491079119e-05,
31
+ "loss": 0.3664,
32
+ "step": 1807
33
+ },
34
+ {
35
+ "epoch": 1.1000553709856036,
36
+ "eval_accuracy": 0.987546699875467,
37
+ "eval_loss": 0.04163224250078201,
38
+ "eval_runtime": 197.9873,
39
+ "eval_samples_per_second": 4.056,
40
+ "eval_steps_per_second": 2.03,
41
+ "step": 1807
42
+ },
43
+ {
44
+ "epoch": 2.1,
45
+ "grad_norm": 0.02027997560799122,
46
+ "learning_rate": 3.8901193552356345e-05,
47
+ "loss": 0.0959,
48
+ "step": 2710
49
+ },
50
+ {
51
+ "epoch": 2.1,
52
+ "eval_accuracy": 0.9912826899128269,
53
+ "eval_loss": 0.03394479304552078,
54
+ "eval_runtime": 180.9016,
55
+ "eval_samples_per_second": 4.439,
56
+ "eval_steps_per_second": 2.222,
57
+ "step": 2710
58
+ },
59
+ {
60
+ "epoch": 3.1000553709856034,
61
+ "grad_norm": 0.0597417987883091,
62
+ "learning_rate": 3.334563799680079e-05,
63
+ "loss": 0.0512,
64
+ "step": 3614
65
+ },
66
+ {
67
+ "epoch": 3.1000553709856034,
68
+ "eval_accuracy": 0.9900373599003736,
69
+ "eval_loss": 0.04761755093932152,
70
+ "eval_runtime": 175.2499,
71
+ "eval_samples_per_second": 4.582,
72
+ "eval_steps_per_second": 2.294,
73
+ "step": 3614
74
+ },
75
+ {
76
+ "epoch": 4.1,
77
+ "grad_norm": 0.05423130467534065,
78
+ "learning_rate": 2.7790082441245237e-05,
79
+ "loss": 0.0461,
80
+ "step": 4517
81
+ },
82
+ {
83
+ "epoch": 4.1,
84
+ "eval_accuracy": 0.9925280199252802,
85
+ "eval_loss": 0.059469074010849,
86
+ "eval_runtime": 177.8384,
87
+ "eval_samples_per_second": 4.515,
88
+ "eval_steps_per_second": 2.26,
89
+ "step": 4517
90
+ },
91
+ {
92
+ "epoch": 5.100055370985603,
93
+ "grad_norm": 0.006802204996347427,
94
+ "learning_rate": 2.222837455395595e-05,
95
+ "loss": 0.0327,
96
+ "step": 5421
97
+ },
98
+ {
99
+ "epoch": 5.100055370985603,
100
+ "eval_accuracy": 0.9962640099626401,
101
+ "eval_loss": 0.01666378229856491,
102
+ "eval_runtime": 173.0935,
103
+ "eval_samples_per_second": 4.639,
104
+ "eval_steps_per_second": 2.322,
105
+ "step": 5421
106
+ },
107
+ {
108
+ "epoch": 6.1,
109
+ "grad_norm": 0.008433694951236248,
110
+ "learning_rate": 1.667897133013412e-05,
111
+ "loss": 0.0247,
112
+ "step": 6324
113
+ },
114
+ {
115
+ "epoch": 6.1,
116
+ "eval_accuracy": 0.9950186799501868,
117
+ "eval_loss": 0.03342529386281967,
118
+ "eval_runtime": 173.3206,
119
+ "eval_samples_per_second": 4.633,
120
+ "eval_steps_per_second": 2.319,
121
+ "step": 6324
122
+ },
123
+ {
124
+ "epoch": 7.100055370985603,
125
+ "grad_norm": 0.001822773483581841,
126
+ "learning_rate": 1.1117263442844839e-05,
127
+ "loss": 0.0252,
128
+ "step": 7228
129
+ },
130
+ {
131
+ "epoch": 7.100055370985603,
132
+ "eval_accuracy": 0.9962640099626401,
133
+ "eval_loss": 0.01384973619133234,
134
+ "eval_runtime": 176.745,
135
+ "eval_samples_per_second": 4.543,
136
+ "eval_steps_per_second": 2.274,
137
+ "step": 7228
138
+ },
139
+ {
140
+ "epoch": 8.1,
141
+ "grad_norm": 0.0026014591567218304,
142
+ "learning_rate": 5.561707887289283e-06,
143
+ "loss": 0.012,
144
+ "step": 8131
145
+ },
146
+ {
147
+ "epoch": 8.1,
148
+ "eval_accuracy": 0.9975093399750934,
149
+ "eval_loss": 0.01134779118001461,
150
+ "eval_runtime": 172.4781,
151
+ "eval_samples_per_second": 4.656,
152
+ "eval_steps_per_second": 2.331,
153
+ "step": 8131
154
+ },
155
+ {
156
+ "epoch": 9.099501661129569,
157
+ "grad_norm": 0.001307124737650156,
158
+ "learning_rate": 3.076165866863542e-08,
159
+ "loss": 0.0002,
160
+ "step": 9030
161
+ },
162
+ {
163
+ "epoch": 9.099501661129569,
164
+ "eval_accuracy": 0.9975093399750934,
165
+ "eval_loss": 0.01189158670604229,
166
+ "eval_runtime": 186.4028,
167
+ "eval_samples_per_second": 4.308,
168
+ "eval_steps_per_second": 2.157,
169
+ "step": 9030
170
+ },
171
+ {
172
+ "epoch": 9.099501661129569,
173
+ "step": 9030,
174
+ "total_flos": 9.006226676074414e+19,
175
+ "train_loss": 0.8929806149349128,
176
+ "train_runtime": 19165.2548,
177
+ "train_samples_per_second": 3.769,
178
+ "train_steps_per_second": 0.471
179
+ }
180
+ ],
181
+ "logging_steps": 500,
182
+ "max_steps": 9030,
183
+ "num_input_tokens_seen": 0,
184
+ "num_train_epochs": 9223372036854775807,
185
+ "save_steps": 500,
186
+ "stateful_callbacks": {
187
+ "EarlyStoppingCallback": {
188
+ "args": {
189
+ "early_stopping_patience": 5,
190
+ "early_stopping_threshold": 0.0
191
+ },
192
+ "attributes": {
193
+ "early_stopping_patience_counter": 1
194
+ }
195
+ },
196
+ "TrainerControl": {
197
+ "args": {
198
+ "should_epoch_stop": false,
199
+ "should_evaluate": false,
200
+ "should_log": false,
201
+ "should_save": true,
202
+ "should_training_stop": true
203
+ },
204
+ "attributes": {}
205
+ }
206
+ },
207
+ "total_flos": 9.006226676074414e+19,
208
+ "train_batch_size": 2,
209
+ "trial_name": null,
210
+ "trial_params": null
211
+ }