SamagraDataGov commited on
Commit
c2f7088
·
verified ·
1 Parent(s): 3c10269

Training in progress, step 40

Browse files
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0d7e37a95520f41735f4e81928cc4a409afc0daed4fe386710ac7f2ac7ad438
3
  size 151099494
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76b38665cb612c63735d262bdbf1373e3aa13465ce6f2dc69e1dc1b9ddef2dbe
3
  size 151099494
runs/Jun29_09-39-32_bharatsahaiyak-test/events.out.tfevents.1719664557.bharatsahaiyak-test.1854617.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a35c59c21e20f207a78d90e2992718b8b9e35524d07f59d44c0c1ab3cb5f5fdb
3
+ size 6396
trainer_state.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "best_metric": 28.246205733558178,
3
- "best_model_checkpoint": "./whisper-tiny-hindi2_test/checkpoint-120",
4
  "epoch": 6.349206349206349,
5
  "eval_steps": 40,
6
  "global_step": 200,
@@ -10,127 +10,127 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.6349206349206349,
13
- "grad_norm": 5.915639877319336,
14
  "learning_rate": 3.75e-05,
15
- "loss": 0.3473,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 1.2698412698412698,
20
- "grad_norm": 3.9101755619049072,
21
  "learning_rate": 3.75e-05,
22
- "loss": 0.2574,
23
  "step": 40
24
  },
25
  {
26
  "epoch": 1.2698412698412698,
27
- "eval_loss": 0.18214347958564758,
28
- "eval_runtime": 56.4889,
29
- "eval_samples_per_second": 1.77,
30
- "eval_steps_per_second": 1.77,
31
- "eval_wer": 29.848229342327148,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 1.9047619047619047,
36
- "grad_norm": 4.230419635772705,
37
  "learning_rate": 3.75e-05,
38
- "loss": 0.1771,
39
  "step": 60
40
  },
41
  {
42
  "epoch": 2.5396825396825395,
43
- "grad_norm": 3.738790273666382,
44
  "learning_rate": 3.75e-05,
45
- "loss": 0.1075,
46
  "step": 80
47
  },
48
  {
49
  "epoch": 2.5396825396825395,
50
- "eval_loss": 0.18291151523590088,
51
- "eval_runtime": 56.5637,
52
- "eval_samples_per_second": 1.768,
53
- "eval_steps_per_second": 1.768,
54
- "eval_wer": 28.920741989881954,
55
  "step": 80
56
  },
57
  {
58
  "epoch": 3.1746031746031744,
59
- "grad_norm": 2.266835927963257,
60
  "learning_rate": 3.75e-05,
61
- "loss": 0.0932,
62
  "step": 100
63
  },
64
  {
65
  "epoch": 3.8095238095238093,
66
- "grad_norm": 2.579704523086548,
67
  "learning_rate": 3.75e-05,
68
- "loss": 0.0513,
69
  "step": 120
70
  },
71
  {
72
  "epoch": 3.8095238095238093,
73
- "eval_loss": 0.18522793054580688,
74
- "eval_runtime": 56.6106,
75
- "eval_samples_per_second": 1.766,
76
- "eval_steps_per_second": 1.766,
77
- "eval_wer": 28.246205733558178,
78
  "step": 120
79
  },
80
  {
81
  "epoch": 4.444444444444445,
82
- "grad_norm": 2.1815927028656006,
83
  "learning_rate": 3.75e-05,
84
- "loss": 0.0365,
85
  "step": 140
86
  },
87
  {
88
  "epoch": 5.079365079365079,
89
- "grad_norm": 2.4605093002319336,
90
  "learning_rate": 3.75e-05,
91
- "loss": 0.0283,
92
  "step": 160
93
  },
94
  {
95
  "epoch": 5.079365079365079,
96
- "eval_loss": 0.19923055171966553,
97
- "eval_runtime": 56.6871,
98
- "eval_samples_per_second": 1.764,
99
- "eval_steps_per_second": 1.764,
100
- "eval_wer": 29.51096121416526,
101
  "step": 160
102
  },
103
  {
104
  "epoch": 5.714285714285714,
105
- "grad_norm": 1.38411283493042,
106
  "learning_rate": 3.75e-05,
107
- "loss": 0.0176,
108
  "step": 180
109
  },
110
  {
111
  "epoch": 6.349206349206349,
112
- "grad_norm": 0.7200319766998291,
113
  "learning_rate": 3.75e-05,
114
- "loss": 0.0178,
115
  "step": 200
116
  },
117
  {
118
  "epoch": 6.349206349206349,
119
- "eval_loss": 0.21108920872211456,
120
- "eval_runtime": 56.6862,
121
- "eval_samples_per_second": 1.764,
122
- "eval_steps_per_second": 1.764,
123
- "eval_wer": 28.499156829679595,
124
  "step": 200
125
  },
126
  {
127
  "epoch": 6.349206349206349,
128
  "step": 200,
129
  "total_flos": 1.5637915828224e+17,
130
- "train_loss": 0.11340499997138977,
131
- "train_runtime": 1142.2335,
132
- "train_samples_per_second": 5.603,
133
- "train_steps_per_second": 0.175
134
  }
135
  ],
136
  "logging_steps": 20,
 
1
  {
2
+ "best_metric": 25.88532883642496,
3
+ "best_model_checkpoint": "./whisper-tiny-hindi2_test/checkpoint-40",
4
  "epoch": 6.349206349206349,
5
  "eval_steps": 40,
6
  "global_step": 200,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.6349206349206349,
13
+ "grad_norm": 5.399875164031982,
14
  "learning_rate": 3.75e-05,
15
+ "loss": 0.3295,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 1.2698412698412698,
20
+ "grad_norm": 3.3471007347106934,
21
  "learning_rate": 3.75e-05,
22
+ "loss": 0.2449,
23
  "step": 40
24
  },
25
  {
26
  "epoch": 1.2698412698412698,
27
+ "eval_loss": 0.15969730913639069,
28
+ "eval_runtime": 56.6547,
29
+ "eval_samples_per_second": 1.765,
30
+ "eval_steps_per_second": 1.765,
31
+ "eval_wer": 25.88532883642496,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 1.9047619047619047,
36
+ "grad_norm": 3.7166926860809326,
37
  "learning_rate": 3.75e-05,
38
+ "loss": 0.1628,
39
  "step": 60
40
  },
41
  {
42
  "epoch": 2.5396825396825395,
43
+ "grad_norm": 2.8016231060028076,
44
  "learning_rate": 3.75e-05,
45
+ "loss": 0.1002,
46
  "step": 80
47
  },
48
  {
49
  "epoch": 2.5396825396825395,
50
+ "eval_loss": 0.16435778141021729,
51
+ "eval_runtime": 56.8979,
52
+ "eval_samples_per_second": 1.758,
53
+ "eval_steps_per_second": 1.758,
54
+ "eval_wer": 26.644182124789207,
55
  "step": 80
56
  },
57
  {
58
  "epoch": 3.1746031746031744,
59
+ "grad_norm": 2.115607976913452,
60
  "learning_rate": 3.75e-05,
61
+ "loss": 0.0833,
62
  "step": 100
63
  },
64
  {
65
  "epoch": 3.8095238095238093,
66
+ "grad_norm": 2.412848711013794,
67
  "learning_rate": 3.75e-05,
68
+ "loss": 0.0492,
69
  "step": 120
70
  },
71
  {
72
  "epoch": 3.8095238095238093,
73
+ "eval_loss": 0.1730262041091919,
74
+ "eval_runtime": 56.4319,
75
+ "eval_samples_per_second": 1.772,
76
+ "eval_steps_per_second": 1.772,
77
+ "eval_wer": 27.065767284991573,
78
  "step": 120
79
  },
80
  {
81
  "epoch": 4.444444444444445,
82
+ "grad_norm": 1.6271615028381348,
83
  "learning_rate": 3.75e-05,
84
+ "loss": 0.0323,
85
  "step": 140
86
  },
87
  {
88
  "epoch": 5.079365079365079,
89
+ "grad_norm": 1.2262483835220337,
90
  "learning_rate": 3.75e-05,
91
+ "loss": 0.0269,
92
  "step": 160
93
  },
94
  {
95
  "epoch": 5.079365079365079,
96
+ "eval_loss": 0.1800723522901535,
97
+ "eval_runtime": 57.7815,
98
+ "eval_samples_per_second": 1.731,
99
+ "eval_steps_per_second": 1.731,
100
+ "eval_wer": 26.981450252951095,
101
  "step": 160
102
  },
103
  {
104
  "epoch": 5.714285714285714,
105
+ "grad_norm": 1.4037569761276245,
106
  "learning_rate": 3.75e-05,
107
+ "loss": 0.0162,
108
  "step": 180
109
  },
110
  {
111
  "epoch": 6.349206349206349,
112
+ "grad_norm": 1.1472747325897217,
113
  "learning_rate": 3.75e-05,
114
+ "loss": 0.0156,
115
  "step": 200
116
  },
117
  {
118
  "epoch": 6.349206349206349,
119
+ "eval_loss": 0.20425428450107574,
120
+ "eval_runtime": 56.7961,
121
+ "eval_samples_per_second": 1.761,
122
+ "eval_steps_per_second": 1.761,
123
+ "eval_wer": 27.65598650927487,
124
  "step": 200
125
  },
126
  {
127
  "epoch": 6.349206349206349,
128
  "step": 200,
129
  "total_flos": 1.5637915828224e+17,
130
+ "train_loss": 0.10607707306742668,
131
+ "train_runtime": 1147.3926,
132
+ "train_samples_per_second": 5.578,
133
+ "train_steps_per_second": 0.174
134
  }
135
  ],
136
  "logging_steps": 20,