youssefedweqd commited on
Commit
22fbc96
·
verified ·
1 Parent(s): 1685348

Training in progress, step 17300

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e471649b89e42bc2f52b00fa86e5bc8195f050f765012f941e595cea8fb3d1d8
3
  size 161533160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a21b464576fcc6c816576b3df5f2f696626ccf8a5cc44dcdd7bab64a60ab0a41
3
  size 161533160
last-checkpoint/adapter_config.json CHANGED
@@ -25,12 +25,12 @@
25
  "revision": null,
26
  "target_modules": [
27
  "q_proj",
28
- "up_proj",
29
- "v_proj",
30
- "gate_proj",
31
  "o_proj",
 
 
 
32
  "k_proj",
33
- "down_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
25
  "revision": null,
26
  "target_modules": [
27
  "q_proj",
 
 
 
28
  "o_proj",
29
+ "gate_proj",
30
+ "down_proj",
31
+ "v_proj",
32
  "k_proj",
33
+ "up_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e471649b89e42bc2f52b00fa86e5bc8195f050f765012f941e595cea8fb3d1d8
3
  size 161533160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f91e3c5f21cf129551bcf7c69bf4d8d16a4ef7f114f3eb09634873a47bacc359
3
  size 161533160
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48a00454de2307c62f3e23d9d447344a70f908c21b8439ac5a93b7f1b4b353b4
3
  size 323292202
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4014d31deb4cce624d0816b7ea019aa4bc828b0109dcbd09f7797bdcac508304
3
  size 323292202
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2bafc1cd3a53197e322d5b1b19353ebaacf1997b784f3897ca35f2fc51e51e3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6863575074205cdae5bd9a3add89cf69a3951c4f077fb3f56f28f0fe9457c14f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 2.847682119205298,
6
  "eval_steps": 100,
7
- "global_step": 17200,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2336,118 +2336,6 @@
2336
  "learning_rate": 2.7228170036058153e-06,
2337
  "loss": 0.2617,
2338
  "step": 16400
2339
- },
2340
- {
2341
- "epoch": 2.7235099337748343,
2342
- "grad_norm": 1.3112074136734009,
2343
- "learning_rate": 2.5682530093579626e-06,
2344
- "loss": 0.2751,
2345
- "step": 16450
2346
- },
2347
- {
2348
- "epoch": 2.7317880794701987,
2349
- "grad_norm": 1.3641598224639893,
2350
- "learning_rate": 2.418089536248769e-06,
2351
- "loss": 0.2741,
2352
- "step": 16500
2353
- },
2354
- {
2355
- "epoch": 2.7400662251655628,
2356
- "grad_norm": 1.2237251996994019,
2357
- "learning_rate": 2.2723405158236455e-06,
2358
- "loss": 0.2772,
2359
- "step": 16550
2360
- },
2361
- {
2362
- "epoch": 2.748344370860927,
2363
- "grad_norm": 1.171391248703003,
2364
- "learning_rate": 2.1310194700733697e-06,
2365
- "loss": 0.2715,
2366
- "step": 16600
2367
- },
2368
- {
2369
- "epoch": 2.756622516556291,
2370
- "grad_norm": 1.3004522323608398,
2371
- "learning_rate": 1.9941395101795655e-06,
2372
- "loss": 0.2712,
2373
- "step": 16650
2374
- },
2375
- {
2376
- "epoch": 2.7649006622516556,
2377
- "grad_norm": 1.0566428899765015,
2378
- "learning_rate": 1.861713335298282e-06,
2379
- "loss": 0.2758,
2380
- "step": 16700
2381
- },
2382
- {
2383
- "epoch": 2.77317880794702,
2384
- "grad_norm": 1.5244101285934448,
2385
- "learning_rate": 1.7337532313818294e-06,
2386
- "loss": 0.2786,
2387
- "step": 16750
2388
- },
2389
- {
2390
- "epoch": 2.781456953642384,
2391
- "grad_norm": 1.2633907794952393,
2392
- "learning_rate": 1.6102710700389656e-06,
2393
- "loss": 0.272,
2394
- "step": 16800
2395
- },
2396
- {
2397
- "epoch": 2.789735099337748,
2398
- "grad_norm": 1.2915066480636597,
2399
- "learning_rate": 1.4912783074334469e-06,
2400
- "loss": 0.2839,
2401
- "step": 16850
2402
- },
2403
- {
2404
- "epoch": 2.7980132450331126,
2405
- "grad_norm": 1.2640457153320312,
2406
- "learning_rate": 1.3767859832212016e-06,
2407
- "loss": 0.2756,
2408
- "step": 16900
2409
- },
2410
- {
2411
- "epoch": 2.806291390728477,
2412
- "grad_norm": 1.3004051446914673,
2413
- "learning_rate": 1.2668047195261379e-06,
2414
- "loss": 0.2747,
2415
- "step": 16950
2416
- },
2417
- {
2418
- "epoch": 2.814569536423841,
2419
- "grad_norm": 1.3884724378585815,
2420
- "learning_rate": 1.1613447199546158e-06,
2421
- "loss": 0.2707,
2422
- "step": 17000
2423
- },
2424
- {
2425
- "epoch": 2.8228476821192054,
2426
- "grad_norm": 1.416150689125061,
2427
- "learning_rate": 1.0604157686488313e-06,
2428
- "loss": 0.2702,
2429
- "step": 17050
2430
- },
2431
- {
2432
- "epoch": 2.8311258278145695,
2433
- "grad_norm": 1.3139573335647583,
2434
- "learning_rate": 9.640272293790998e-07,
2435
- "loss": 0.2721,
2436
- "step": 17100
2437
- },
2438
- {
2439
- "epoch": 2.839403973509934,
2440
- "grad_norm": 1.2322123050689697,
2441
- "learning_rate": 8.721880446750708e-07,
2442
- "loss": 0.2734,
2443
- "step": 17150
2444
- },
2445
- {
2446
- "epoch": 2.847682119205298,
2447
- "grad_norm": 1.3550212383270264,
2448
- "learning_rate": 7.849067349961381e-07,
2449
- "loss": 0.2757,
2450
- "step": 17200
2451
  }
2452
  ],
2453
  "logging_steps": 50,
@@ -2467,7 +2355,7 @@
2467
  "attributes": {}
2468
  }
2469
  },
2470
- "total_flos": 1.350085670904791e+17,
2471
  "train_batch_size": 1,
2472
  "trial_name": null,
2473
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 2.7152317880794703,
6
  "eval_steps": 100,
7
+ "global_step": 16400,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2336
  "learning_rate": 2.7228170036058153e-06,
2337
  "loss": 0.2617,
2338
  "step": 16400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2339
  }
2340
  ],
2341
  "logging_steps": 50,
 
2355
  "attributes": {}
2356
  }
2357
  },
2358
+ "total_flos": 1.2874605314388787e+17,
2359
  "train_batch_size": 1,
2360
  "trial_name": null,
2361
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e16cd32295896d2f57f1f19dbc49310a6ac6c29a2ec738cf4fb0c11d34db588b
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e7fd7045b4a15373c4d905f915edbb1436664635b44fe1df1bfa95de5188043
3
  size 5752
trainer_log.jsonl CHANGED
@@ -14,3 +14,5 @@
14
  {"current_steps": 17100, "total_steps": 18120, "loss": 0.2721, "lr": 9.640272293790998e-07, "epoch": 2.8311258278145695, "percentage": 94.37, "elapsed_time": "1:14:28", "remaining_time": "0:04:26"}
15
  {"current_steps": 17150, "total_steps": 18120, "loss": 0.2734, "lr": 8.721880446750708e-07, "epoch": 2.839403973509934, "percentage": 94.65, "elapsed_time": "1:19:42", "remaining_time": "0:04:30"}
16
  {"current_steps": 17200, "total_steps": 18120, "loss": 0.2757, "lr": 7.849067349961381e-07, "epoch": 2.847682119205298, "percentage": 94.92, "elapsed_time": "1:24:42", "remaining_time": "0:04:31"}
 
 
 
14
  {"current_steps": 17100, "total_steps": 18120, "loss": 0.2721, "lr": 9.640272293790998e-07, "epoch": 2.8311258278145695, "percentage": 94.37, "elapsed_time": "1:14:28", "remaining_time": "0:04:26"}
15
  {"current_steps": 17150, "total_steps": 18120, "loss": 0.2734, "lr": 8.721880446750708e-07, "epoch": 2.839403973509934, "percentage": 94.65, "elapsed_time": "1:19:42", "remaining_time": "0:04:30"}
16
  {"current_steps": 17200, "total_steps": 18120, "loss": 0.2757, "lr": 7.849067349961381e-07, "epoch": 2.847682119205298, "percentage": 94.92, "elapsed_time": "1:24:42", "remaining_time": "0:04:31"}
17
+ {"current_steps": 17250, "total_steps": 18120, "loss": 0.274, "lr": 7.02191397940899e-07, "epoch": 2.8559602649006623, "percentage": 95.2, "elapsed_time": "1:29:52", "remaining_time": "0:04:31"}
18
+ {"current_steps": 17300, "total_steps": 18120, "loss": 0.2721, "lr": 6.240497074959162e-07, "epoch": 2.8642384105960264, "percentage": 95.47, "elapsed_time": "1:34:46", "remaining_time": "0:04:29"}