Myat-Oo-Swe commited on
Commit
dcce8c9
·
verified ·
1 Parent(s): d7ed253

Training in progress, step 1500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3edcd0c450189261c882afd547ab8aa2ed8d8d69f6514bdd036f696631191522
3
  size 62969640
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:239e6ee34c627288f07da635c7607c26cb1658248cb4216138582c969c562c53
3
  size 62969640
last-checkpoint/adapter_model/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3edcd0c450189261c882afd547ab8aa2ed8d8d69f6514bdd036f696631191522
3
  size 62969640
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:239e6ee34c627288f07da635c7607c26cb1658248cb4216138582c969c562c53
3
  size 62969640
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c5dcfe80745e49d3dda94c607cdd0c6937698a1eee8f03ccc8d98c943d8860e
3
  size 84101370
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b57dc2a75271934b9272c6e364dbf8843d89c17ee905195eb7b2c5eaf32f6c3c
3
  size 84101370
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:658a00b68b67cee9ed9e480f0e8e586986ce85df79a4a11fa9577da999c85123
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:018d3489ed13a5f5303fe87bc05ee3bd7e42fab42cae25b01d59fef9cd4802ee
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d8fdcd0311eba9854fff738038ed4c1a269832665b4d88ba4e4e3d02a1a7e0e
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2a4b6e0ef05ca249f2d6b7f9f3ad1dff81e55842a962df795bb9740c17c8e92
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97a379574c9663f8f6b1e4e8f5372c7153e8ea3e03443c673cf8a63a64a20ef3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:416c83758baa78b1c3e5334d8425b4107684ccd443376b2af239f14f10f93e92
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 1000,
3
- "best_metric": 0.6822749972343445,
4
- "best_model_checkpoint": "./whisper-large-v3-lt1/checkpoint-1000",
5
- "epoch": 0.5,
6
  "eval_steps": 500,
7
- "global_step": 1000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -94,6 +94,49 @@
94
  "eval_samples_per_second": 2.504,
95
  "eval_steps_per_second": 0.313,
96
  "step": 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
97
  }
98
  ],
99
  "logging_steps": 100,
@@ -113,7 +156,7 @@
113
  "attributes": {}
114
  }
115
  },
116
- "total_flos": 5.493979348992e+19,
117
  "train_batch_size": 16,
118
  "trial_name": null,
119
  "trial_params": null
 
1
  {
2
+ "best_global_step": 1500,
3
+ "best_metric": 0.6280742287635803,
4
+ "best_model_checkpoint": "./whisper-large-v3-lt1/checkpoint-1500",
5
+ "epoch": 0.75,
6
  "eval_steps": 500,
7
+ "global_step": 1500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
94
  "eval_samples_per_second": 2.504,
95
  "eval_steps_per_second": 0.313,
96
  "step": 1000
97
+ },
98
+ {
99
+ "epoch": 0.55,
100
+ "grad_norm": 1.2514197826385498,
101
+ "learning_rate": 4.7421052631578954e-06,
102
+ "loss": 0.6689,
103
+ "step": 1100
104
+ },
105
+ {
106
+ "epoch": 0.6,
107
+ "grad_norm": 2.271310806274414,
108
+ "learning_rate": 4.215789473684211e-06,
109
+ "loss": 0.6526,
110
+ "step": 1200
111
+ },
112
+ {
113
+ "epoch": 0.65,
114
+ "grad_norm": 2.1460328102111816,
115
+ "learning_rate": 3.6894736842105265e-06,
116
+ "loss": 0.6299,
117
+ "step": 1300
118
+ },
119
+ {
120
+ "epoch": 0.7,
121
+ "grad_norm": 2.67410945892334,
122
+ "learning_rate": 3.1631578947368424e-06,
123
+ "loss": 0.6143,
124
+ "step": 1400
125
+ },
126
+ {
127
+ "epoch": 0.75,
128
+ "grad_norm": 1.5648741722106934,
129
+ "learning_rate": 2.6368421052631584e-06,
130
+ "loss": 0.6632,
131
+ "step": 1500
132
+ },
133
+ {
134
+ "epoch": 0.75,
135
+ "eval_loss": 0.6280742287635803,
136
+ "eval_runtime": 1163.7066,
137
+ "eval_samples_per_second": 2.509,
138
+ "eval_steps_per_second": 0.314,
139
+ "step": 1500
140
  }
141
  ],
142
  "logging_steps": 100,
 
156
  "attributes": {}
157
  }
158
  },
159
+ "total_flos": 8.240969023488e+19,
160
  "train_batch_size": 16,
161
  "trial_name": null,
162
  "trial_params": null