SamagraDataGov commited on
Commit
1ff7eeb
·
verified ·
1 Parent(s): 1bb2263

Training in progress, step 40

Browse files
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28d00ca3bfb780e4dc72d46545fa78a6ba7b4ff64ff7fca8ceca22f284d3373c
3
  size 151099494
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec43cf6eef1f5aa761a37e943c182de501d2bc55b4d9d8424a011f69ffa73a1e
3
  size 151099494
runs/Jun29_09-39-32_bharatsahaiyak-test/events.out.tfevents.1719657502.bharatsahaiyak-test.1854617.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80ae0d0376797cbf383586bd9fc07951e66d4f11e270768f2186a245c1741852
3
+ size 6396
trainer_state.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "best_metric": 41.31534569983137,
3
- "best_model_checkpoint": "./whisper-tiny-hindi2_test/checkpoint-160",
4
  "epoch": 6.349206349206349,
5
  "eval_steps": 40,
6
  "global_step": 200,
@@ -10,127 +10,127 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.6349206349206349,
13
- "grad_norm": 5.291396141052246,
14
  "learning_rate": 3.75e-05,
15
- "loss": 0.486,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 1.2698412698412698,
20
- "grad_norm": 4.965972423553467,
21
  "learning_rate": 3.75e-05,
22
- "loss": 0.3869,
23
  "step": 40
24
  },
25
  {
26
  "epoch": 1.2698412698412698,
27
- "eval_loss": 0.35420987010002136,
28
- "eval_runtime": 54.9702,
29
- "eval_samples_per_second": 1.819,
30
- "eval_steps_per_second": 1.819,
31
- "eval_wer": 46.964586846543,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 1.9047619047619047,
36
- "grad_norm": 5.148536682128906,
37
  "learning_rate": 3.75e-05,
38
- "loss": 0.3043,
39
  "step": 60
40
  },
41
  {
42
  "epoch": 2.5396825396825395,
43
- "grad_norm": 4.049417972564697,
44
  "learning_rate": 3.75e-05,
45
- "loss": 0.2215,
46
  "step": 80
47
  },
48
  {
49
  "epoch": 2.5396825396825395,
50
- "eval_loss": 0.3464047908782959,
51
- "eval_runtime": 55.3552,
52
- "eval_samples_per_second": 1.807,
53
- "eval_steps_per_second": 1.807,
54
- "eval_wer": 45.27824620573356,
55
  "step": 80
56
  },
57
  {
58
  "epoch": 3.1746031746031744,
59
- "grad_norm": 3.4034841060638428,
60
  "learning_rate": 3.75e-05,
61
- "loss": 0.1888,
62
  "step": 100
63
  },
64
  {
65
  "epoch": 3.8095238095238093,
66
- "grad_norm": 4.5857930183410645,
67
  "learning_rate": 3.75e-05,
68
- "loss": 0.1394,
69
  "step": 120
70
  },
71
  {
72
  "epoch": 3.8095238095238093,
73
- "eval_loss": 0.34948885440826416,
74
- "eval_runtime": 54.9671,
75
- "eval_samples_per_second": 1.819,
76
- "eval_steps_per_second": 1.819,
77
- "eval_wer": 43.00168634064081,
78
  "step": 120
79
  },
80
  {
81
  "epoch": 4.444444444444445,
82
- "grad_norm": 3.3065757751464844,
83
  "learning_rate": 3.75e-05,
84
- "loss": 0.0954,
85
  "step": 140
86
  },
87
  {
88
  "epoch": 5.079365079365079,
89
- "grad_norm": 2.241292953491211,
90
  "learning_rate": 3.75e-05,
91
- "loss": 0.0857,
92
  "step": 160
93
  },
94
  {
95
  "epoch": 5.079365079365079,
96
- "eval_loss": 0.37522462010383606,
97
- "eval_runtime": 55.4528,
98
- "eval_samples_per_second": 1.803,
99
- "eval_steps_per_second": 1.803,
100
- "eval_wer": 41.31534569983137,
101
  "step": 160
102
  },
103
  {
104
  "epoch": 5.714285714285714,
105
- "grad_norm": 2.4962286949157715,
106
  "learning_rate": 3.75e-05,
107
- "loss": 0.055,
108
  "step": 180
109
  },
110
  {
111
  "epoch": 6.349206349206349,
112
- "grad_norm": 2.652437210083008,
113
  "learning_rate": 3.75e-05,
114
- "loss": 0.0434,
115
  "step": 200
116
  },
117
  {
118
  "epoch": 6.349206349206349,
119
- "eval_loss": 0.40884342789649963,
120
- "eval_runtime": 55.4835,
121
- "eval_samples_per_second": 1.802,
122
- "eval_steps_per_second": 1.802,
123
- "eval_wer": 43.76053962900506,
124
  "step": 200
125
  },
126
  {
127
  "epoch": 6.349206349206349,
128
  "step": 200,
129
  "total_flos": 1.5637915828224e+17,
130
- "train_loss": 0.2006339970231056,
131
- "train_runtime": 1120.101,
132
- "train_samples_per_second": 5.714,
133
- "train_steps_per_second": 0.179
134
  }
135
  ],
136
  "logging_steps": 20,
 
1
  {
2
+ "best_metric": 37.77403035413153,
3
+ "best_model_checkpoint": "./whisper-tiny-hindi2_test/checkpoint-80",
4
  "epoch": 6.349206349206349,
5
  "eval_steps": 40,
6
  "global_step": 200,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.6349206349206349,
13
+ "grad_norm": 6.332822322845459,
14
  "learning_rate": 3.75e-05,
15
+ "loss": 0.4561,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 1.2698412698412698,
20
+ "grad_norm": 4.458395957946777,
21
  "learning_rate": 3.75e-05,
22
+ "loss": 0.3384,
23
  "step": 40
24
  },
25
  {
26
  "epoch": 1.2698412698412698,
27
+ "eval_loss": 0.300150990486145,
28
+ "eval_runtime": 55.4195,
29
+ "eval_samples_per_second": 1.804,
30
+ "eval_steps_per_second": 1.804,
31
+ "eval_wer": 39.62900505902192,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 1.9047619047619047,
36
+ "grad_norm": 3.924806594848633,
37
  "learning_rate": 3.75e-05,
38
+ "loss": 0.2515,
39
  "step": 60
40
  },
41
  {
42
  "epoch": 2.5396825396825395,
43
+ "grad_norm": 3.3504045009613037,
44
  "learning_rate": 3.75e-05,
45
+ "loss": 0.1811,
46
  "step": 80
47
  },
48
  {
49
  "epoch": 2.5396825396825395,
50
+ "eval_loss": 0.2907748520374298,
51
+ "eval_runtime": 55.3083,
52
+ "eval_samples_per_second": 1.808,
53
+ "eval_steps_per_second": 1.808,
54
+ "eval_wer": 37.77403035413153,
55
  "step": 80
56
  },
57
  {
58
  "epoch": 3.1746031746031744,
59
+ "grad_norm": 3.752678155899048,
60
  "learning_rate": 3.75e-05,
61
+ "loss": 0.1512,
62
  "step": 100
63
  },
64
  {
65
  "epoch": 3.8095238095238093,
66
+ "grad_norm": 3.3088107109069824,
67
  "learning_rate": 3.75e-05,
68
+ "loss": 0.0989,
69
  "step": 120
70
  },
71
  {
72
  "epoch": 3.8095238095238093,
73
+ "eval_loss": 0.3248939514160156,
74
+ "eval_runtime": 55.1423,
75
+ "eval_samples_per_second": 1.813,
76
+ "eval_steps_per_second": 1.813,
77
+ "eval_wer": 39.79763912310287,
78
  "step": 120
79
  },
80
  {
81
  "epoch": 4.444444444444445,
82
+ "grad_norm": 3.1989450454711914,
83
  "learning_rate": 3.75e-05,
84
+ "loss": 0.0738,
85
  "step": 140
86
  },
87
  {
88
  "epoch": 5.079365079365079,
89
+ "grad_norm": 2.052924871444702,
90
  "learning_rate": 3.75e-05,
91
+ "loss": 0.0623,
92
  "step": 160
93
  },
94
  {
95
  "epoch": 5.079365079365079,
96
+ "eval_loss": 0.33577826619148254,
97
+ "eval_runtime": 56.8542,
98
+ "eval_samples_per_second": 1.759,
99
+ "eval_steps_per_second": 1.759,
100
+ "eval_wer": 45.362563237774026,
101
  "step": 160
102
  },
103
  {
104
  "epoch": 5.714285714285714,
105
+ "grad_norm": 1.9767564535140991,
106
  "learning_rate": 3.75e-05,
107
+ "loss": 0.0377,
108
  "step": 180
109
  },
110
  {
111
  "epoch": 6.349206349206349,
112
+ "grad_norm": 3.543325185775757,
113
  "learning_rate": 3.75e-05,
114
+ "loss": 0.0344,
115
  "step": 200
116
  },
117
  {
118
  "epoch": 6.349206349206349,
119
+ "eval_loss": 0.3560102880001068,
120
+ "eval_runtime": 55.9482,
121
+ "eval_samples_per_second": 1.787,
122
+ "eval_steps_per_second": 1.787,
123
+ "eval_wer": 40.80944350758853,
124
  "step": 200
125
  },
126
  {
127
  "epoch": 6.349206349206349,
128
  "step": 200,
129
  "total_flos": 1.5637915828224e+17,
130
+ "train_loss": 0.16853376895189284,
131
+ "train_runtime": 1128.2867,
132
+ "train_samples_per_second": 5.672,
133
+ "train_steps_per_second": 0.177
134
  }
135
  ],
136
  "logging_steps": 20,