Myat-Oo-Swe commited on
Commit
ca4560d
·
verified ·
1 Parent(s): c595b4a

Training in progress, step 1000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d76ac46219d8d6464b4275617d85ce9aea730b11a6e0e9fca3eebd3de90fda63
3
  size 62969640
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3edcd0c450189261c882afd547ab8aa2ed8d8d69f6514bdd036f696631191522
3
  size 62969640
last-checkpoint/adapter_model/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d76ac46219d8d6464b4275617d85ce9aea730b11a6e0e9fca3eebd3de90fda63
3
  size 62969640
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3edcd0c450189261c882afd547ab8aa2ed8d8d69f6514bdd036f696631191522
3
  size 62969640
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9cb556ea2c9b9b10fca0d9e95927bd0f59def6539e0e5b6fe039e3407785d505
3
  size 84101370
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c5dcfe80745e49d3dda94c607cdd0c6937698a1eee8f03ccc8d98c943d8860e
3
  size 84101370
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b27a72f606fe83c400f5cf18fccfbeb6b3aeb5bd692f12d1975d2e9fa01952a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:658a00b68b67cee9ed9e480f0e8e586986ce85df79a4a11fa9577da999c85123
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18b984273ea2d45b7ffb1d047bb359d93111e41fcad70d16a1b453fd38f72636
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d8fdcd0311eba9854fff738038ed4c1a269832665b4d88ba4e4e3d02a1a7e0e
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c6c3e39e65b038f3305b10302a421e7d2279cc0b6df6c74ce7b9308a8a5cda9
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97a379574c9663f8f6b1e4e8f5372c7153e8ea3e03443c673cf8a63a64a20ef3
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 500,
3
- "best_metric": 0.8364643454551697,
4
- "best_model_checkpoint": "./whisper-large-v3-lt1/checkpoint-500",
5
- "epoch": 0.25,
6
  "eval_steps": 500,
7
- "global_step": 500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -51,6 +51,49 @@
51
  "eval_samples_per_second": 2.503,
52
  "eval_steps_per_second": 0.313,
53
  "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
54
  }
55
  ],
56
  "logging_steps": 100,
@@ -70,7 +113,7 @@
70
  "attributes": {}
71
  }
72
  },
73
- "total_flos": 2.746989674496e+19,
74
  "train_batch_size": 16,
75
  "trial_name": null,
76
  "trial_params": null
 
1
  {
2
+ "best_global_step": 1000,
3
+ "best_metric": 0.6822749972343445,
4
+ "best_model_checkpoint": "./whisper-large-v3-lt1/checkpoint-1000",
5
+ "epoch": 0.5,
6
  "eval_steps": 500,
7
+ "global_step": 1000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
51
  "eval_samples_per_second": 2.503,
52
  "eval_steps_per_second": 0.313,
53
  "step": 500
54
+ },
55
+ {
56
+ "epoch": 0.3,
57
+ "grad_norm": 1.2657830715179443,
58
+ "learning_rate": 7.373684210526316e-06,
59
+ "loss": 0.796,
60
+ "step": 600
61
+ },
62
+ {
63
+ "epoch": 0.35,
64
+ "grad_norm": 1.17739737033844,
65
+ "learning_rate": 6.8473684210526325e-06,
66
+ "loss": 0.8109,
67
+ "step": 700
68
+ },
69
+ {
70
+ "epoch": 0.4,
71
+ "grad_norm": 1.271132230758667,
72
+ "learning_rate": 6.321052631578948e-06,
73
+ "loss": 0.7474,
74
+ "step": 800
75
+ },
76
+ {
77
+ "epoch": 0.45,
78
+ "grad_norm": 1.3513257503509521,
79
+ "learning_rate": 5.794736842105264e-06,
80
+ "loss": 0.7134,
81
+ "step": 900
82
+ },
83
+ {
84
+ "epoch": 0.5,
85
+ "grad_norm": 1.3304741382598877,
86
+ "learning_rate": 5.268421052631579e-06,
87
+ "loss": 0.6847,
88
+ "step": 1000
89
+ },
90
+ {
91
+ "epoch": 0.5,
92
+ "eval_loss": 0.6822749972343445,
93
+ "eval_runtime": 1166.2313,
94
+ "eval_samples_per_second": 2.504,
95
+ "eval_steps_per_second": 0.313,
96
+ "step": 1000
97
  }
98
  ],
99
  "logging_steps": 100,
 
113
  "attributes": {}
114
  }
115
  },
116
+ "total_flos": 5.493979348992e+19,
117
  "train_batch_size": 16,
118
  "trial_name": null,
119
  "trial_params": null