madeofajala commited on
Commit
d79c8dc
·
verified ·
1 Parent(s): 1be2780

Training in progress, step 2850, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -29,13 +29,13 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
 
 
32
  "down_proj",
33
- "q_proj",
34
- "gate_proj",
35
  "v_proj",
36
- "k_proj",
37
- "o_proj",
38
- "up_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "up_proj",
33
+ "o_proj",
34
+ "k_proj",
35
  "down_proj",
 
 
36
  "v_proj",
37
+ "q_proj",
38
+ "gate_proj"
 
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f729200728ba44a9ee48ef9f7a746ee4013df5b4441295deb59e7e533620a072
3
  size 108113968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd9069c95306b127720930a77fad3324cb0d7eda4a839db90d5605d41852db31
3
  size 108113968
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:971c9d1c5b74e903eb9d3f94fe8af4db0d58ee1f9669348434fb9761d8371e7b
3
  size 57081771
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22c04e45f62499567b0d0bf6dc6fe8b59e32b7921c27ac5943ee035968aa590d
3
  size 57081771
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b81ad22ca863c3bf9e4ae89b44de94db92351a68d9ae0bd607ab062743a9371
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ed08c1881bd90abbd6a4f09b7d8dcdcd76153d3d15f906bdc29b7fedf5425cb
3
  size 1465
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.8096875895672112,
6
  "eval_steps": 300,
7
- "global_step": 2825,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1138,12 +1138,22 @@
1138
  "mean_token_accuracy": 0.9419060254096985,
1139
  "num_tokens": 250323.0,
1140
  "step": 2825
 
 
 
 
 
 
 
 
 
 
1141
  }
1142
  ],
1143
  "logging_steps": 25,
1144
  "max_steps": 3489,
1145
  "num_input_tokens_seen": 0,
1146
- "num_train_epochs": 1,
1147
  "save_steps": 25,
1148
  "stateful_callbacks": {
1149
  "TrainerControl": {
@@ -1157,7 +1167,7 @@
1157
  "attributes": {}
1158
  }
1159
  },
1160
- "total_flos": 2.7417012377326387e+17,
1161
  "train_batch_size": 4,
1162
  "trial_name": null,
1163
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 1.6334193178561192,
6
  "eval_steps": 300,
7
+ "global_step": 2850,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1138
  "mean_token_accuracy": 0.9419060254096985,
1139
  "num_tokens": 250323.0,
1140
  "step": 2825
1141
+ },
1142
+ {
1143
+ "entropy": 0.15991646617650987,
1144
+ "epoch": 1.6334193178561192,
1145
+ "grad_norm": 0.19363631308078766,
1146
+ "learning_rate": 0.0002,
1147
+ "loss": 0.15416499137878417,
1148
+ "mean_token_accuracy": 0.9481097102165222,
1149
+ "num_tokens": 82145.0,
1150
+ "step": 2850
1151
  }
1152
  ],
1153
  "logging_steps": 25,
1154
  "max_steps": 3489,
1155
  "num_input_tokens_seen": 0,
1156
+ "num_train_epochs": 2,
1157
  "save_steps": 25,
1158
  "stateful_callbacks": {
1159
  "TrainerControl": {
 
1167
  "attributes": {}
1168
  }
1169
  },
1170
+ "total_flos": 2.7903573973558886e+17,
1171
  "train_batch_size": 4,
1172
  "trial_name": null,
1173
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b6b8cd0e3f39da2d16e5b5e0bab780af893d21cb8f98534cd79dde7079e4d98
3
  size 5649
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96134eafdc3eaeac8127a27f3b984cecea62d6c6dc170b55ad5610e26cab65b1
3
  size 5649