SamagraDataGov commited on
Commit
ef24e8f
·
verified ·
1 Parent(s): aafdf38

Training in progress, step 40

Browse files
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6993d2eb56920db28a25a0241d646d0cc3d87980efa46bdd0408aa766dcdcb94
3
  size 151099494
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0d7e37a95520f41735f4e81928cc4a409afc0daed4fe386710ac7f2ac7ad438
3
  size 151099494
runs/Jun29_09-39-32_bharatsahaiyak-test/events.out.tfevents.1719663371.bharatsahaiyak-test.1854617.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c33d876b5bbb51ab64eb5c09b0e89150b69c5e644a90af2cafd4ecb9216df90
3
+ size 6396
trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 29.34232715008432,
3
  "best_model_checkpoint": "./whisper-tiny-hindi2_test/checkpoint-120",
4
  "epoch": 6.349206349206349,
5
  "eval_steps": 40,
@@ -10,127 +10,127 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.6349206349206349,
13
- "grad_norm": 6.46198034286499,
14
  "learning_rate": 3.75e-05,
15
- "loss": 0.3431,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 1.2698412698412698,
20
- "grad_norm": 3.6580495834350586,
21
  "learning_rate": 3.75e-05,
22
- "loss": 0.257,
23
  "step": 40
24
  },
25
  {
26
  "epoch": 1.2698412698412698,
27
- "eval_loss": 0.19398407638072968,
28
- "eval_runtime": 55.7217,
29
- "eval_samples_per_second": 1.795,
30
- "eval_steps_per_second": 1.795,
31
- "eval_wer": 30.94435075885329,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 1.9047619047619047,
36
- "grad_norm": 4.121974468231201,
37
  "learning_rate": 3.75e-05,
38
- "loss": 0.1865,
39
  "step": 60
40
  },
41
  {
42
  "epoch": 2.5396825396825395,
43
- "grad_norm": 3.0512349605560303,
44
  "learning_rate": 3.75e-05,
45
- "loss": 0.1177,
46
  "step": 80
47
  },
48
  {
49
  "epoch": 2.5396825396825395,
50
- "eval_loss": 0.1945016235113144,
51
- "eval_runtime": 55.3505,
52
- "eval_samples_per_second": 1.807,
53
- "eval_steps_per_second": 1.807,
54
- "eval_wer": 31.197301854974707,
55
  "step": 80
56
  },
57
  {
58
  "epoch": 3.1746031746031744,
59
- "grad_norm": 2.3403635025024414,
60
  "learning_rate": 3.75e-05,
61
- "loss": 0.0903,
62
  "step": 100
63
  },
64
  {
65
  "epoch": 3.8095238095238093,
66
- "grad_norm": 2.4136340618133545,
67
  "learning_rate": 3.75e-05,
68
- "loss": 0.0569,
69
  "step": 120
70
  },
71
  {
72
  "epoch": 3.8095238095238093,
73
- "eval_loss": 0.20778781175613403,
74
- "eval_runtime": 55.9933,
75
- "eval_samples_per_second": 1.786,
76
- "eval_steps_per_second": 1.786,
77
- "eval_wer": 29.34232715008432,
78
  "step": 120
79
  },
80
  {
81
  "epoch": 4.444444444444445,
82
- "grad_norm": 1.9897055625915527,
83
  "learning_rate": 3.75e-05,
84
- "loss": 0.0412,
85
  "step": 140
86
  },
87
  {
88
  "epoch": 5.079365079365079,
89
- "grad_norm": 1.213258981704712,
90
  "learning_rate": 3.75e-05,
91
- "loss": 0.0313,
92
  "step": 160
93
  },
94
  {
95
  "epoch": 5.079365079365079,
96
- "eval_loss": 0.2349608838558197,
97
- "eval_runtime": 56.0059,
98
- "eval_samples_per_second": 1.786,
99
- "eval_steps_per_second": 1.786,
100
- "eval_wer": 31.02866779089376,
101
  "step": 160
102
  },
103
  {
104
  "epoch": 5.714285714285714,
105
- "grad_norm": 1.2564868927001953,
106
  "learning_rate": 3.75e-05,
107
- "loss": 0.0198,
108
  "step": 180
109
  },
110
  {
111
  "epoch": 6.349206349206349,
112
- "grad_norm": 1.6000959873199463,
113
  "learning_rate": 3.75e-05,
114
- "loss": 0.0186,
115
  "step": 200
116
  },
117
  {
118
  "epoch": 6.349206349206349,
119
- "eval_loss": 0.2441704422235489,
120
- "eval_runtime": 55.4225,
121
- "eval_samples_per_second": 1.804,
122
- "eval_steps_per_second": 1.804,
123
- "eval_wer": 30.6070826306914,
124
  "step": 200
125
  },
126
  {
127
  "epoch": 6.349206349206349,
128
  "step": 200,
129
  "total_flos": 1.5637915828224e+17,
130
- "train_loss": 0.11623679384589196,
131
- "train_runtime": 1122.2532,
132
- "train_samples_per_second": 5.703,
133
- "train_steps_per_second": 0.178
134
  }
135
  ],
136
  "logging_steps": 20,
 
1
  {
2
+ "best_metric": 28.246205733558178,
3
  "best_model_checkpoint": "./whisper-tiny-hindi2_test/checkpoint-120",
4
  "epoch": 6.349206349206349,
5
  "eval_steps": 40,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.6349206349206349,
13
+ "grad_norm": 5.915639877319336,
14
  "learning_rate": 3.75e-05,
15
+ "loss": 0.3473,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 1.2698412698412698,
20
+ "grad_norm": 3.9101755619049072,
21
  "learning_rate": 3.75e-05,
22
+ "loss": 0.2574,
23
  "step": 40
24
  },
25
  {
26
  "epoch": 1.2698412698412698,
27
+ "eval_loss": 0.18214347958564758,
28
+ "eval_runtime": 56.4889,
29
+ "eval_samples_per_second": 1.77,
30
+ "eval_steps_per_second": 1.77,
31
+ "eval_wer": 29.848229342327148,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 1.9047619047619047,
36
+ "grad_norm": 4.230419635772705,
37
  "learning_rate": 3.75e-05,
38
+ "loss": 0.1771,
39
  "step": 60
40
  },
41
  {
42
  "epoch": 2.5396825396825395,
43
+ "grad_norm": 3.738790273666382,
44
  "learning_rate": 3.75e-05,
45
+ "loss": 0.1075,
46
  "step": 80
47
  },
48
  {
49
  "epoch": 2.5396825396825395,
50
+ "eval_loss": 0.18291151523590088,
51
+ "eval_runtime": 56.5637,
52
+ "eval_samples_per_second": 1.768,
53
+ "eval_steps_per_second": 1.768,
54
+ "eval_wer": 28.920741989881954,
55
  "step": 80
56
  },
57
  {
58
  "epoch": 3.1746031746031744,
59
+ "grad_norm": 2.266835927963257,
60
  "learning_rate": 3.75e-05,
61
+ "loss": 0.0932,
62
  "step": 100
63
  },
64
  {
65
  "epoch": 3.8095238095238093,
66
+ "grad_norm": 2.579704523086548,
67
  "learning_rate": 3.75e-05,
68
+ "loss": 0.0513,
69
  "step": 120
70
  },
71
  {
72
  "epoch": 3.8095238095238093,
73
+ "eval_loss": 0.18522793054580688,
74
+ "eval_runtime": 56.6106,
75
+ "eval_samples_per_second": 1.766,
76
+ "eval_steps_per_second": 1.766,
77
+ "eval_wer": 28.246205733558178,
78
  "step": 120
79
  },
80
  {
81
  "epoch": 4.444444444444445,
82
+ "grad_norm": 2.1815927028656006,
83
  "learning_rate": 3.75e-05,
84
+ "loss": 0.0365,
85
  "step": 140
86
  },
87
  {
88
  "epoch": 5.079365079365079,
89
+ "grad_norm": 2.4605093002319336,
90
  "learning_rate": 3.75e-05,
91
+ "loss": 0.0283,
92
  "step": 160
93
  },
94
  {
95
  "epoch": 5.079365079365079,
96
+ "eval_loss": 0.19923055171966553,
97
+ "eval_runtime": 56.6871,
98
+ "eval_samples_per_second": 1.764,
99
+ "eval_steps_per_second": 1.764,
100
+ "eval_wer": 29.51096121416526,
101
  "step": 160
102
  },
103
  {
104
  "epoch": 5.714285714285714,
105
+ "grad_norm": 1.38411283493042,
106
  "learning_rate": 3.75e-05,
107
+ "loss": 0.0176,
108
  "step": 180
109
  },
110
  {
111
  "epoch": 6.349206349206349,
112
+ "grad_norm": 0.7200319766998291,
113
  "learning_rate": 3.75e-05,
114
+ "loss": 0.0178,
115
  "step": 200
116
  },
117
  {
118
  "epoch": 6.349206349206349,
119
+ "eval_loss": 0.21108920872211456,
120
+ "eval_runtime": 56.6862,
121
+ "eval_samples_per_second": 1.764,
122
+ "eval_steps_per_second": 1.764,
123
+ "eval_wer": 28.499156829679595,
124
  "step": 200
125
  },
126
  {
127
  "epoch": 6.349206349206349,
128
  "step": 200,
129
  "total_flos": 1.5637915828224e+17,
130
+ "train_loss": 0.11340499997138977,
131
+ "train_runtime": 1142.2335,
132
+ "train_samples_per_second": 5.603,
133
+ "train_steps_per_second": 0.175
134
  }
135
  ],
136
  "logging_steps": 20,