SamagraDataGov commited on
Commit
3c656d2
·
verified ·
1 Parent(s): 21f76c5

Training in progress, step 40

Browse files
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e83c4fe02f5392ab3ed8a461043b0b685b305a863f3a18e3a1fca823e66a06ca
3
  size 151099494
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea5038f27370638d492fddc58c333f1164ffa5af2486a0669818856adcdd8fe6
3
  size 151099494
runs/Jun29_09-39-32_bharatsahaiyak-test/events.out.tfevents.1719656331.bharatsahaiyak-test.1854617.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15a929009023974b064687f87b39150a992f59c732c03358d85440603f7e558a
3
+ size 6396
trainer_state.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "best_metric": 50.0,
3
- "best_model_checkpoint": "./whisper-tiny-hindi2_test/checkpoint-120",
4
  "epoch": 6.349206349206349,
5
  "eval_steps": 40,
6
  "global_step": 200,
@@ -10,127 +10,127 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.6349206349206349,
13
- "grad_norm": 8.679319381713867,
14
  "learning_rate": 3.75e-05,
15
- "loss": 1.6908,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 1.2698412698412698,
20
- "grad_norm": 7.228819370269775,
21
  "learning_rate": 3.75e-05,
22
- "loss": 0.6584,
23
  "step": 40
24
  },
25
  {
26
  "epoch": 1.2698412698412698,
27
- "eval_loss": 0.593582272529602,
28
- "eval_runtime": 54.5955,
29
- "eval_samples_per_second": 1.832,
30
- "eval_steps_per_second": 1.832,
31
- "eval_wer": 65.00843170320405,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 1.9047619047619047,
36
- "grad_norm": 6.050016403198242,
37
  "learning_rate": 3.75e-05,
38
- "loss": 0.4786,
39
  "step": 60
40
  },
41
  {
42
  "epoch": 2.5396825396825395,
43
- "grad_norm": 5.063595294952393,
44
  "learning_rate": 3.75e-05,
45
- "loss": 0.3575,
46
  "step": 80
47
  },
48
  {
49
  "epoch": 2.5396825396825395,
50
- "eval_loss": 0.46842461824417114,
51
- "eval_runtime": 55.6599,
52
- "eval_samples_per_second": 1.797,
53
- "eval_steps_per_second": 1.797,
54
- "eval_wer": 55.31197301854974,
55
  "step": 80
56
  },
57
  {
58
  "epoch": 3.1746031746031744,
59
- "grad_norm": 4.799503326416016,
60
  "learning_rate": 3.75e-05,
61
- "loss": 0.2994,
62
  "step": 100
63
  },
64
  {
65
  "epoch": 3.8095238095238093,
66
- "grad_norm": 5.387985706329346,
67
  "learning_rate": 3.75e-05,
68
- "loss": 0.2254,
69
  "step": 120
70
  },
71
  {
72
  "epoch": 3.8095238095238093,
73
- "eval_loss": 0.4344378709793091,
74
- "eval_runtime": 54.7109,
75
- "eval_samples_per_second": 1.828,
76
- "eval_steps_per_second": 1.828,
77
- "eval_wer": 50.0,
78
  "step": 120
79
  },
80
  {
81
  "epoch": 4.444444444444445,
82
- "grad_norm": 4.451031684875488,
83
  "learning_rate": 3.75e-05,
84
- "loss": 0.1752,
85
  "step": 140
86
  },
87
  {
88
  "epoch": 5.079365079365079,
89
- "grad_norm": 4.960638046264648,
90
  "learning_rate": 3.75e-05,
91
- "loss": 0.1486,
92
  "step": 160
93
  },
94
  {
95
  "epoch": 5.079365079365079,
96
- "eval_loss": 0.4561767578125,
97
- "eval_runtime": 54.4171,
98
- "eval_samples_per_second": 1.838,
99
- "eval_steps_per_second": 1.838,
100
- "eval_wer": 52.1079258010118,
101
  "step": 160
102
  },
103
  {
104
  "epoch": 5.714285714285714,
105
- "grad_norm": 3.8416409492492676,
106
  "learning_rate": 3.75e-05,
107
- "loss": 0.1021,
108
  "step": 180
109
  },
110
  {
111
  "epoch": 6.349206349206349,
112
- "grad_norm": 3.137124538421631,
113
  "learning_rate": 3.75e-05,
114
- "loss": 0.0868,
115
  "step": 200
116
  },
117
  {
118
  "epoch": 6.349206349206349,
119
- "eval_loss": 0.46873587369918823,
120
- "eval_runtime": 54.2397,
121
- "eval_samples_per_second": 1.844,
122
- "eval_steps_per_second": 1.844,
123
- "eval_wer": 50.5902192242833,
124
  "step": 200
125
  },
126
  {
127
  "epoch": 6.349206349206349,
128
  "step": 200,
129
  "total_flos": 1.5637915828224e+17,
130
- "train_loss": 0.4222871881723404,
131
- "train_runtime": 1142.3513,
132
- "train_samples_per_second": 5.602,
133
- "train_steps_per_second": 0.175
134
  }
135
  ],
136
  "logging_steps": 20,
 
1
  {
2
+ "best_metric": 41.31534569983137,
3
+ "best_model_checkpoint": "./whisper-tiny-hindi2_test/checkpoint-160",
4
  "epoch": 6.349206349206349,
5
  "eval_steps": 40,
6
  "global_step": 200,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.6349206349206349,
13
+ "grad_norm": 5.291396141052246,
14
  "learning_rate": 3.75e-05,
15
+ "loss": 0.486,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 1.2698412698412698,
20
+ "grad_norm": 4.965972423553467,
21
  "learning_rate": 3.75e-05,
22
+ "loss": 0.3869,
23
  "step": 40
24
  },
25
  {
26
  "epoch": 1.2698412698412698,
27
+ "eval_loss": 0.35420987010002136,
28
+ "eval_runtime": 54.9702,
29
+ "eval_samples_per_second": 1.819,
30
+ "eval_steps_per_second": 1.819,
31
+ "eval_wer": 46.964586846543,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 1.9047619047619047,
36
+ "grad_norm": 5.148536682128906,
37
  "learning_rate": 3.75e-05,
38
+ "loss": 0.3043,
39
  "step": 60
40
  },
41
  {
42
  "epoch": 2.5396825396825395,
43
+ "grad_norm": 4.049417972564697,
44
  "learning_rate": 3.75e-05,
45
+ "loss": 0.2215,
46
  "step": 80
47
  },
48
  {
49
  "epoch": 2.5396825396825395,
50
+ "eval_loss": 0.3464047908782959,
51
+ "eval_runtime": 55.3552,
52
+ "eval_samples_per_second": 1.807,
53
+ "eval_steps_per_second": 1.807,
54
+ "eval_wer": 45.27824620573356,
55
  "step": 80
56
  },
57
  {
58
  "epoch": 3.1746031746031744,
59
+ "grad_norm": 3.4034841060638428,
60
  "learning_rate": 3.75e-05,
61
+ "loss": 0.1888,
62
  "step": 100
63
  },
64
  {
65
  "epoch": 3.8095238095238093,
66
+ "grad_norm": 4.5857930183410645,
67
  "learning_rate": 3.75e-05,
68
+ "loss": 0.1394,
69
  "step": 120
70
  },
71
  {
72
  "epoch": 3.8095238095238093,
73
+ "eval_loss": 0.34948885440826416,
74
+ "eval_runtime": 54.9671,
75
+ "eval_samples_per_second": 1.819,
76
+ "eval_steps_per_second": 1.819,
77
+ "eval_wer": 43.00168634064081,
78
  "step": 120
79
  },
80
  {
81
  "epoch": 4.444444444444445,
82
+ "grad_norm": 3.3065757751464844,
83
  "learning_rate": 3.75e-05,
84
+ "loss": 0.0954,
85
  "step": 140
86
  },
87
  {
88
  "epoch": 5.079365079365079,
89
+ "grad_norm": 2.241292953491211,
90
  "learning_rate": 3.75e-05,
91
+ "loss": 0.0857,
92
  "step": 160
93
  },
94
  {
95
  "epoch": 5.079365079365079,
96
+ "eval_loss": 0.37522462010383606,
97
+ "eval_runtime": 55.4528,
98
+ "eval_samples_per_second": 1.803,
99
+ "eval_steps_per_second": 1.803,
100
+ "eval_wer": 41.31534569983137,
101
  "step": 160
102
  },
103
  {
104
  "epoch": 5.714285714285714,
105
+ "grad_norm": 2.4962286949157715,
106
  "learning_rate": 3.75e-05,
107
+ "loss": 0.055,
108
  "step": 180
109
  },
110
  {
111
  "epoch": 6.349206349206349,
112
+ "grad_norm": 2.652437210083008,
113
  "learning_rate": 3.75e-05,
114
+ "loss": 0.0434,
115
  "step": 200
116
  },
117
  {
118
  "epoch": 6.349206349206349,
119
+ "eval_loss": 0.40884342789649963,
120
+ "eval_runtime": 55.4835,
121
+ "eval_samples_per_second": 1.802,
122
+ "eval_steps_per_second": 1.802,
123
+ "eval_wer": 43.76053962900506,
124
  "step": 200
125
  },
126
  {
127
  "epoch": 6.349206349206349,
128
  "step": 200,
129
  "total_flos": 1.5637915828224e+17,
130
+ "train_loss": 0.2006339970231056,
131
+ "train_runtime": 1120.101,
132
+ "train_samples_per_second": 5.714,
133
+ "train_steps_per_second": 0.179
134
  }
135
  ],
136
  "logging_steps": 20,