youssefedweqd commited on
Commit
3651bb4
·
verified ·
1 Parent(s): 0ac8689

Training in progress, step 16600

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:41c469e4040081e0ecfe377eeb4ac213060f49bafda35052223ed0946061838f
3
  size 161533160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fa435505cd53f72a810061ae0179d098c329f5cd5a93585597b00b60e2b0852
3
  size 161533160
last-checkpoint/adapter_config.json CHANGED
@@ -25,12 +25,12 @@
25
  "revision": null,
26
  "target_modules": [
27
  "q_proj",
28
- "up_proj",
29
- "v_proj",
30
- "gate_proj",
31
  "o_proj",
 
 
 
32
  "k_proj",
33
- "down_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
25
  "revision": null,
26
  "target_modules": [
27
  "q_proj",
 
 
 
28
  "o_proj",
29
+ "gate_proj",
30
+ "down_proj",
31
+ "v_proj",
32
  "k_proj",
33
+ "up_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:41c469e4040081e0ecfe377eeb4ac213060f49bafda35052223ed0946061838f
3
  size 161533160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f91e3c5f21cf129551bcf7c69bf4d8d16a4ef7f114f3eb09634873a47bacc359
3
  size 161533160
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c768b1437290a19e8d89a3a9d42b5710c409a98d8a94298c5825f8294a4fffe
3
  size 323292202
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4014d31deb4cce624d0816b7ea019aa4bc828b0109dcbd09f7797bdcac508304
3
  size 323292202
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fbc25b7042f4401d966ff059f93c12cc7772c87508517760f6980236a9daf83b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6863575074205cdae5bd9a3add89cf69a3951c4f077fb3f56f28f0fe9457c14f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 2.7317880794701987,
6
  "eval_steps": 100,
7
- "global_step": 16500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2336,20 +2336,6 @@
2336
  "learning_rate": 2.7228170036058153e-06,
2337
  "loss": 0.2617,
2338
  "step": 16400
2339
- },
2340
- {
2341
- "epoch": 2.7235099337748343,
2342
- "grad_norm": 1.3112074136734009,
2343
- "learning_rate": 2.5682530093579626e-06,
2344
- "loss": 0.2751,
2345
- "step": 16450
2346
- },
2347
- {
2348
- "epoch": 2.7317880794701987,
2349
- "grad_norm": 1.3641598224639893,
2350
- "learning_rate": 2.418089536248769e-06,
2351
- "loss": 0.2741,
2352
- "step": 16500
2353
  }
2354
  ],
2355
  "logging_steps": 50,
@@ -2369,7 +2355,7 @@
2369
  "attributes": {}
2370
  }
2371
  },
2372
- "total_flos": 1.295039378793431e+17,
2373
  "train_batch_size": 1,
2374
  "trial_name": null,
2375
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 2.7152317880794703,
6
  "eval_steps": 100,
7
+ "global_step": 16400,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2336
  "learning_rate": 2.7228170036058153e-06,
2337
  "loss": 0.2617,
2338
  "step": 16400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2339
  }
2340
  ],
2341
  "logging_steps": 50,
 
2355
  "attributes": {}
2356
  }
2357
  },
2358
+ "total_flos": 1.2874605314388787e+17,
2359
  "train_batch_size": 1,
2360
  "trial_name": null,
2361
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e16cd32295896d2f57f1f19dbc49310a6ac6c29a2ec738cf4fb0c11d34db588b
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e7fd7045b4a15373c4d905f915edbb1436664635b44fe1df1bfa95de5188043
3
  size 5752
trainer_log.jsonl CHANGED
@@ -1,2 +1,4 @@
1
  {"current_steps": 16450, "total_steps": 18120, "loss": 0.2751, "lr": 2.5682530093579626e-06, "epoch": 2.7235099337748343, "percentage": 90.78, "elapsed_time": "0:05:13", "remaining_time": "0:00:31"}
2
  {"current_steps": 16500, "total_steps": 18120, "loss": 0.2741, "lr": 2.418089536248769e-06, "epoch": 2.7317880794701987, "percentage": 91.06, "elapsed_time": "0:10:04", "remaining_time": "0:00:59"}
 
 
 
1
  {"current_steps": 16450, "total_steps": 18120, "loss": 0.2751, "lr": 2.5682530093579626e-06, "epoch": 2.7235099337748343, "percentage": 90.78, "elapsed_time": "0:05:13", "remaining_time": "0:00:31"}
2
  {"current_steps": 16500, "total_steps": 18120, "loss": 0.2741, "lr": 2.418089536248769e-06, "epoch": 2.7317880794701987, "percentage": 91.06, "elapsed_time": "0:10:04", "remaining_time": "0:00:59"}
3
+ {"current_steps": 16550, "total_steps": 18120, "loss": 0.2772, "lr": 2.2723405158236455e-06, "epoch": 2.7400662251655628, "percentage": 91.34, "elapsed_time": "0:15:23", "remaining_time": "0:01:27"}
4
+ {"current_steps": 16600, "total_steps": 18120, "loss": 0.2715, "lr": 2.1310194700733697e-06, "epoch": 2.748344370860927, "percentage": 91.61, "elapsed_time": "0:20:25", "remaining_time": "0:01:52"}