Shawon16 commited on
Commit
b9386c8
·
verified ·
1 Parent(s): 298b82a

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ confusion_matrix_kfold_fold_5.tiff filter=lfs diff=lfs merge=lfs -text
37
+ confusion_matrix_test_fold_5.tiff filter=lfs diff=lfs merge=lfs -text
38
+ confusion_matrix_trainfold_fold_5.tiff filter=lfs diff=lfs merge=lfs -text
39
+ long_tail_performance_vs_class_frequency_kfold_fold_5.tiff filter=lfs diff=lfs merge=lfs -text
40
+ long_tail_performance_vs_class_frequency_test_fold_5.tiff filter=lfs diff=lfs merge=lfs -text
41
+ long_tail_performance_vs_class_frequency_trainfold_fold_5.tiff filter=lfs diff=lfs merge=lfs -text
42
+ per_class_accuracy_kfold_fold_5.tiff filter=lfs diff=lfs merge=lfs -text
43
+ per_class_accuracy_test_fold_5.tiff filter=lfs diff=lfs merge=lfs -text
44
+ per_class_accuracy_trainfold_fold_5.tiff filter=lfs diff=lfs merge=lfs -text
confusion_matrix_kfold_fold_5.tiff ADDED

Git LFS Details

  • SHA256: a216540ecf63e316c667be8b5a3ca2caa3487ca01b3979d7146256e7e92c2c15
  • Pointer size: 133 Bytes
  • Size of remote file: 13 MB
confusion_matrix_test_fold_5.tiff ADDED

Git LFS Details

  • SHA256: 020d8bebcd668924b0fd75c8f6fd5b80b89dfda568b76ae5ccc922e4ede0adfa
  • Pointer size: 133 Bytes
  • Size of remote file: 12.9 MB
confusion_matrix_trainfold_fold_5.tiff ADDED

Git LFS Details

  • SHA256: cfbcc72047ec9cdd9d01492eb9c2741347bd2b6dd99003936b5c8dca046efca8
  • Pointer size: 133 Bytes
  • Size of remote file: 13 MB
long_tail_performance_vs_class_frequency_kfold_fold_5.tiff ADDED

Git LFS Details

  • SHA256: 0a7feffff55a0e868272e7eb5f51f7edda34c0b1d45f48b80d4c0b4215afd302
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
long_tail_performance_vs_class_frequency_test_fold_5.tiff ADDED

Git LFS Details

  • SHA256: 2b8f23d91ba7caf9031db52991940418f91af0718aaa768ba8aaa41972d55259
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
long_tail_performance_vs_class_frequency_trainfold_fold_5.tiff ADDED

Git LFS Details

  • SHA256: d6461ebb9aea214a2616ce16be6aee1a89036115c0dd3ab9ea115c71aa23f566
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_kfold_fold_5.tiff ADDED

Git LFS Details

  • SHA256: 00c5e0be1e1cccab5bfdd59ddf8d3372fd65edb887475b3eabf8330908ffd224
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_test_fold_5.tiff ADDED

Git LFS Details

  • SHA256: f5a66962c66b9d1ad47dbf6b35295456cc4fb2b92c72899f2f3834764c70d3b9
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_accuracy_trainfold_fold_5.tiff ADDED

Git LFS Details

  • SHA256: bd6a20f8f6ad0167c2e6280938d6152817dc800acc9a389053625a1edd648f5f
  • Pointer size: 133 Bytes
  • Size of remote file: 26.5 MB
per_class_results_fold_5.csv ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ fold,class,train_accuracy,train_precision,train_recall,train_f1_score,train_top5_accuracy,val_accuracy,val_precision,val_recall,val_f1_score,val_top5_accuracy,test_accuracy,test_precision,test_recall,test_f1_score,test_top5_accuracy
2
+ 5,cake,0.9920634920634921,1.0,0.9920634920634921,0.9960159362549801,1.0,0.9285714285714286,1.0,0.9285714285714286,0.9629629629629629,1.0,0.47619047619047616,1.0,0.47619047619047616,0.6451612903225806,0.9898119122257053
3
+ 5,aids,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.6666666666666666,0.8,0.6666666666666666,0.7272727272727273,0.9898119122257053
4
+ 5,chikissha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.95,1.0,0.95,0.9743589743589743,0.9898119122257053
5
+ 5,maa,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9523809523809523,1.0,0.975609756097561,0.9898119122257053
6
+ 5,durbol,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
7
+ 5,toothpaste,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
8
+ 5,apartment,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
9
+ 5,chal,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.7083333333333334,1.0,0.7083333333333334,0.8292682926829268,0.9898119122257053
10
+ 5,chadar,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
11
+ 5,bon,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9,0.9473684210526315,0.9,0.9230769230769231,0.9898119122257053
12
+ 5,ac,0.972972972972973,1.0,0.972972972972973,0.9863013698630136,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.7586206896551724,1.0,0.8627450980392157,0.9898119122257053
13
+ 5,konna,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
14
+ 5,baba,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
15
+ 5,cream,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.38095238095238093,1.0,0.38095238095238093,0.5517241379310345,0.9898119122257053
16
+ 5,chips,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
17
+ 5,chacha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9545454545454546,1.0,0.9545454545454546,0.9767441860465116,0.9898119122257053
18
+ 5,dulavai,0.9920634920634921,1.0,0.9920634920634921,0.9960159362549801,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.92,1.0,0.9583333333333334,0.9898119122257053
19
+ 5,balu,1.0,0.96875,1.0,0.9841269841269841,1.0,1.0,0.9333333333333333,1.0,0.9655172413793104,1.0,0.9545454545454546,0.5833333333333334,0.9545454545454546,0.7241379310344828,0.9898119122257053
20
+ 5,baat,1.0,0.9920634920634921,1.0,0.9960159362549801,1.0,1.0,0.9333333333333333,1.0,0.9655172413793104,1.0,0.9545454545454546,0.9545454545454546,0.9545454545454546,0.9545454545454546,0.9898119122257053
21
+ 5,attio,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9090909090909091,0.9090909090909091,0.9090909090909091,0.9090909090909091,0.9898119122257053
22
+ 5,capsule,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
23
+ 5,denadar,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.95,1.0,0.9743589743589743,0.9898119122257053
24
+ 5,biscuts,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9523809523809523,1.0,0.975609756097561,0.9898119122257053
25
+ 5,churi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
26
+ 5,tattha,1.0,1.0,1.0,1.0,1.0,0.8461538461538461,1.0,0.8461538461538461,0.9166666666666666,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
27
+ 5,clip,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
28
+ 5,chiruni,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
29
+ 5,doctor,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
30
+ 5,tv,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
31
+ 5,baandej,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9047619047619048,1.0,0.95,0.9898119122257053
32
+ 5,aam,1.0,1.0,1.0,1.0,1.0,1.0,0.9285714285714286,1.0,0.9629629629629629,1.0,1.0,0.9545454545454546,1.0,0.9767441860465116,0.9898119122257053
33
+ 5,ayna,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
34
+ 5,alu,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.8076923076923077,1.0,0.8936170212765957,0.9898119122257053
35
+ 5,daeitto,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9615384615384616,1.0,0.9803921568627451,0.9898119122257053
36
+ 5,cha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.7272727272727273,1.0,0.7272727272727273,0.8421052631578947,0.9898119122257053
37
+ 5,audio cassette,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
38
+ 5,dongson,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
39
+ 5,balti,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.75,1.0,0.75,0.8571428571428571,0.9898119122257053
40
+ 5,angur,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
41
+ 5,dadi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
42
+ 5,chachi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9090909090909091,0.6451612903225806,0.9090909090909091,0.7547169811320755,0.9898119122257053
43
+ 5,bhai,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
44
+ 5,chosma,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9130434782608695,1.0,0.9130434782608695,0.9545454545454546,0.9898119122257053
45
+ 5,debor,0.9917355371900827,1.0,0.9917355371900827,0.995850622406639,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
46
+ 5,dal,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9545454545454546,0.9545454545454546,0.9545454545454546,0.9545454545454546,0.9898119122257053
47
+ 5,bou,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9130434782608695,1.0,0.9545454545454546,0.9898119122257053
48
+ 5,dada,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9130434782608695,1.0,0.9545454545454546,0.9898119122257053
49
+ 5,anaros,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9130434782608695,1.0,0.9130434782608695,0.9545454545454546,0.9898119122257053
50
+ 5,boroi,1.0,0.9923664122137404,1.0,0.9961685823754789,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9310344827586207,1.0,0.9642857142857143,0.9898119122257053
51
+ 5,chokh utha,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
52
+ 5,aaple,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
53
+ 5,chocolate,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.88,1.0,0.9361702127659575,0.9898119122257053
54
+ 5,tshirt,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9545454545454546,1.0,0.9767441860465116,0.9898119122257053
55
+ 5,bottam,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
56
+ 5,juta,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9,0.9473684210526315,0.9,0.9230769230769231,0.9898119122257053
57
+ 5,tubelight,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
58
+ 5,tupi,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
59
+ 5,dengue,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
60
+ 5,jomoj,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9583333333333334,1.0,0.9583333333333334,0.9787234042553191,0.9898119122257053
61
+ 5,chini,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,1.0,0.9898119122257053
trainer_state.json ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9962640099626401,
3
+ "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/BdSLW60_10_Fold/VideoMAE_Kinetics_fold__4__BdSLW60_SKF/checkpoint-5421",
4
+ "epoch": 9.099501661129569,
5
+ "eval_steps": 500,
6
+ "global_step": 9030,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.1,
13
+ "grad_norm": 49.20745849609375,
14
+ "learning_rate": 4.9889258028792915e-05,
15
+ "loss": 8.2618,
16
+ "step": 903
17
+ },
18
+ {
19
+ "epoch": 0.1,
20
+ "eval_accuracy": 0.9564134495641345,
21
+ "eval_loss": 0.30000391602516174,
22
+ "eval_runtime": 158.3701,
23
+ "eval_samples_per_second": 5.07,
24
+ "eval_steps_per_second": 2.538,
25
+ "step": 903
26
+ },
27
+ {
28
+ "epoch": 1.1000553709856036,
29
+ "grad_norm": 4.823966026306152,
30
+ "learning_rate": 4.44567491079119e-05,
31
+ "loss": 0.3664,
32
+ "step": 1807
33
+ },
34
+ {
35
+ "epoch": 1.1000553709856036,
36
+ "eval_accuracy": 0.9850560398505604,
37
+ "eval_loss": 0.051344048231840134,
38
+ "eval_runtime": 167.1002,
39
+ "eval_samples_per_second": 4.806,
40
+ "eval_steps_per_second": 2.406,
41
+ "step": 1807
42
+ },
43
+ {
44
+ "epoch": 2.1,
45
+ "grad_norm": 0.04141543060541153,
46
+ "learning_rate": 3.8901193552356345e-05,
47
+ "loss": 0.0972,
48
+ "step": 2710
49
+ },
50
+ {
51
+ "epoch": 2.1,
52
+ "eval_accuracy": 0.987546699875467,
53
+ "eval_loss": 0.05115814134478569,
54
+ "eval_runtime": 162.4904,
55
+ "eval_samples_per_second": 4.942,
56
+ "eval_steps_per_second": 2.474,
57
+ "step": 2710
58
+ },
59
+ {
60
+ "epoch": 3.1000553709856034,
61
+ "grad_norm": 0.010718895122408867,
62
+ "learning_rate": 3.334563799680079e-05,
63
+ "loss": 0.0595,
64
+ "step": 3614
65
+ },
66
+ {
67
+ "epoch": 3.1000553709856034,
68
+ "eval_accuracy": 0.9863013698630136,
69
+ "eval_loss": 0.05742684379220009,
70
+ "eval_runtime": 164.5821,
71
+ "eval_samples_per_second": 4.879,
72
+ "eval_steps_per_second": 2.443,
73
+ "step": 3614
74
+ },
75
+ {
76
+ "epoch": 4.1,
77
+ "grad_norm": 0.026031654328107834,
78
+ "learning_rate": 2.7796234772978964e-05,
79
+ "loss": 0.0477,
80
+ "step": 4517
81
+ },
82
+ {
83
+ "epoch": 4.1,
84
+ "eval_accuracy": 0.9950186799501868,
85
+ "eval_loss": 0.02064591273665428,
86
+ "eval_runtime": 176.3267,
87
+ "eval_samples_per_second": 4.554,
88
+ "eval_steps_per_second": 2.28,
89
+ "step": 4517
90
+ },
91
+ {
92
+ "epoch": 5.100055370985603,
93
+ "grad_norm": 0.007739614229649305,
94
+ "learning_rate": 2.2234526885689677e-05,
95
+ "loss": 0.0525,
96
+ "step": 5421
97
+ },
98
+ {
99
+ "epoch": 5.100055370985603,
100
+ "eval_accuracy": 0.9962640099626401,
101
+ "eval_loss": 0.013474504463374615,
102
+ "eval_runtime": 170.3372,
103
+ "eval_samples_per_second": 4.714,
104
+ "eval_steps_per_second": 2.36,
105
+ "step": 5421
106
+ },
107
+ {
108
+ "epoch": 6.1,
109
+ "grad_norm": 0.06612707674503326,
110
+ "learning_rate": 1.668512366186785e-05,
111
+ "loss": 0.0129,
112
+ "step": 6324
113
+ },
114
+ {
115
+ "epoch": 6.1,
116
+ "eval_accuracy": 0.9962640099626401,
117
+ "eval_loss": 0.02480306848883629,
118
+ "eval_runtime": 166.308,
119
+ "eval_samples_per_second": 4.828,
120
+ "eval_steps_per_second": 2.417,
121
+ "step": 6324
122
+ },
123
+ {
124
+ "epoch": 7.100055370985603,
125
+ "grad_norm": 0.0024384052958339453,
126
+ "learning_rate": 1.1123415774578566e-05,
127
+ "loss": 0.0149,
128
+ "step": 7228
129
+ },
130
+ {
131
+ "epoch": 7.100055370985603,
132
+ "eval_accuracy": 0.9950186799501868,
133
+ "eval_loss": 0.016877437010407448,
134
+ "eval_runtime": 167.4524,
135
+ "eval_samples_per_second": 4.795,
136
+ "eval_steps_per_second": 2.401,
137
+ "step": 7228
138
+ },
139
+ {
140
+ "epoch": 8.1,
141
+ "grad_norm": 0.0029954672791063786,
142
+ "learning_rate": 5.574012550756737e-06,
143
+ "loss": 0.0028,
144
+ "step": 8131
145
+ },
146
+ {
147
+ "epoch": 8.1,
148
+ "eval_accuracy": 0.9950186799501868,
149
+ "eval_loss": 0.013714665547013283,
150
+ "eval_runtime": 167.9061,
151
+ "eval_samples_per_second": 4.782,
152
+ "eval_steps_per_second": 2.394,
153
+ "step": 8131
154
+ },
155
+ {
156
+ "epoch": 9.099501661129569,
157
+ "grad_norm": 0.009799107909202576,
158
+ "learning_rate": 4.306632213608958e-08,
159
+ "loss": 0.002,
160
+ "step": 9030
161
+ },
162
+ {
163
+ "epoch": 9.099501661129569,
164
+ "eval_accuracy": 0.9962640099626401,
165
+ "eval_loss": 0.011564984917640686,
166
+ "eval_runtime": 184.4167,
167
+ "eval_samples_per_second": 4.354,
168
+ "eval_steps_per_second": 2.18,
169
+ "step": 9030
170
+ },
171
+ {
172
+ "epoch": 9.099501661129569,
173
+ "step": 9030,
174
+ "total_flos": 9.006226676074414e+19,
175
+ "train_loss": 0.8918199165978965,
176
+ "train_runtime": 19029.9044,
177
+ "train_samples_per_second": 3.796,
178
+ "train_steps_per_second": 0.475
179
+ }
180
+ ],
181
+ "logging_steps": 500,
182
+ "max_steps": 9030,
183
+ "num_input_tokens_seen": 0,
184
+ "num_train_epochs": 9223372036854775807,
185
+ "save_steps": 500,
186
+ "stateful_callbacks": {
187
+ "EarlyStoppingCallback": {
188
+ "args": {
189
+ "early_stopping_patience": 5,
190
+ "early_stopping_threshold": 0.0
191
+ },
192
+ "attributes": {
193
+ "early_stopping_patience_counter": 4
194
+ }
195
+ },
196
+ "TrainerControl": {
197
+ "args": {
198
+ "should_epoch_stop": false,
199
+ "should_evaluate": false,
200
+ "should_log": false,
201
+ "should_save": true,
202
+ "should_training_stop": true
203
+ },
204
+ "attributes": {}
205
+ }
206
+ },
207
+ "total_flos": 9.006226676074414e+19,
208
+ "train_batch_size": 2,
209
+ "trial_name": null,
210
+ "trial_params": null
211
+ }