besimray commited on
Commit
558eca1
·
verified ·
1 Parent(s): 691ed91

Training in progress, step 5, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "gate_proj",
24
- "v_proj",
25
- "down_proj",
26
  "up_proj",
27
  "o_proj",
28
- "k_proj",
29
- "q_proj"
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "up_proj",
24
  "o_proj",
25
+ "down_proj",
26
+ "gate_proj",
27
+ "v_proj",
28
+ "q_proj",
29
+ "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43da2a1eb2790fe62f293e01d8ada013a0d966df4f2ada2660b9a71af9f5002a
3
  size 45118424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3e79621ffc14fb91a39d698499c2cb759cf1caea21cf010b8e80340ac6a5721
3
  size 45118424
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39701fa216f95fd1e0171751e98aa83608c72ac9eb0c890e4ab69484314eeeed
3
  size 23159290
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f6c2ef00fa3cffe1608f2bcd6476e992598cc48a8279a6f80478bfff85bff11
3
  size 23159290
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13893f254074b0dec5e8003565583ed9154abab49baf146f6045b5d48206e343
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:413baf801b6b282f4a4f8a102791a0e43f3596f80253ee4e57b1d684a62cbb52
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bb578e75c11a81e85dda67a691f96ba4793a02960f1409fd3e1511aac873491a
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cda2d7da3ce8a95be6df505b84eace6ccd5aa18ffc6d1bcc9a79572045c8d78f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,130 +1,63 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.21052631578947367,
5
- "eval_steps": 2,
6
- "global_step": 10,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.021052631578947368,
13
- "grad_norm": 0.7498180866241455,
14
  "learning_rate": 2e-05,
15
- "loss": 1.2983,
16
  "step": 1
17
  },
18
  {
19
- "epoch": 0.021052631578947368,
20
- "eval_loss": 1.2585705518722534,
21
- "eval_runtime": 15.0051,
22
- "eval_samples_per_second": 6.664,
23
- "eval_steps_per_second": 0.666,
24
  "step": 1
25
  },
26
  {
27
- "epoch": 0.042105263157894736,
28
- "grad_norm": 0.7510321736335754,
29
  "learning_rate": 4e-05,
30
- "loss": 1.4576,
31
- "step": 2
32
- },
33
- {
34
- "epoch": 0.042105263157894736,
35
- "eval_loss": 1.2596663236618042,
36
- "eval_runtime": 15.033,
37
- "eval_samples_per_second": 6.652,
38
- "eval_steps_per_second": 0.665,
39
  "step": 2
40
  },
41
  {
42
- "epoch": 0.06315789473684211,
43
- "grad_norm": 0.6777486801147461,
44
  "learning_rate": 6e-05,
45
- "loss": 1.2738,
46
  "step": 3
47
  },
48
  {
49
- "epoch": 0.08421052631578947,
50
- "grad_norm": 0.7081015706062317,
51
- "learning_rate": 8e-05,
52
- "loss": 1.4211,
53
- "step": 4
 
54
  },
55
  {
56
- "epoch": 0.08421052631578947,
57
- "eval_loss": 1.2397140264511108,
58
- "eval_runtime": 15.0876,
59
- "eval_samples_per_second": 6.628,
60
- "eval_steps_per_second": 0.663,
61
  "step": 4
62
  },
63
  {
64
- "epoch": 0.10526315789473684,
65
- "grad_norm": 0.5160530805587769,
66
  "learning_rate": 0.0001,
67
- "loss": 1.2373,
68
  "step": 5
69
- },
70
- {
71
- "epoch": 0.12631578947368421,
72
- "grad_norm": 0.5380682349205017,
73
- "learning_rate": 0.00012,
74
- "loss": 1.3893,
75
- "step": 6
76
- },
77
- {
78
- "epoch": 0.12631578947368421,
79
- "eval_loss": 1.2157351970672607,
80
- "eval_runtime": 15.0706,
81
- "eval_samples_per_second": 6.635,
82
- "eval_steps_per_second": 0.664,
83
- "step": 6
84
- },
85
- {
86
- "epoch": 0.14736842105263157,
87
- "grad_norm": 0.6841205358505249,
88
- "learning_rate": 0.00014,
89
- "loss": 1.3356,
90
- "step": 7
91
- },
92
- {
93
- "epoch": 0.16842105263157894,
94
- "grad_norm": 0.4957394599914551,
95
- "learning_rate": 0.00016,
96
- "loss": 1.2216,
97
- "step": 8
98
- },
99
- {
100
- "epoch": 0.16842105263157894,
101
- "eval_loss": 1.2087069749832153,
102
- "eval_runtime": 15.1596,
103
- "eval_samples_per_second": 6.596,
104
- "eval_steps_per_second": 0.66,
105
- "step": 8
106
- },
107
- {
108
- "epoch": 0.18947368421052632,
109
- "grad_norm": 0.5728506445884705,
110
- "learning_rate": 0.00018,
111
- "loss": 1.2016,
112
- "step": 9
113
- },
114
- {
115
- "epoch": 0.21052631578947367,
116
- "grad_norm": 0.4859888255596161,
117
- "learning_rate": 0.0002,
118
- "loss": 1.3764,
119
- "step": 10
120
- },
121
- {
122
- "epoch": 0.21052631578947367,
123
- "eval_loss": 1.1915773153305054,
124
- "eval_runtime": 15.0938,
125
- "eval_samples_per_second": 6.625,
126
- "eval_steps_per_second": 0.663,
127
- "step": 10
128
  }
129
  ],
130
  "logging_steps": 1,
@@ -139,13 +72,13 @@
139
  "should_evaluate": false,
140
  "should_log": false,
141
  "should_save": true,
142
- "should_training_stop": true
143
  },
144
  "attributes": {}
145
  }
146
  },
147
- "total_flos": 9677225695641600.0,
148
- "train_batch_size": 10,
149
  "trial_name": null,
150
  "trial_params": null
151
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.05263157894736842,
5
+ "eval_steps": 3,
6
+ "global_step": 5,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.010526315789473684,
13
+ "grad_norm": 0.7775679230690002,
14
  "learning_rate": 2e-05,
15
+ "loss": 1.3487,
16
  "step": 1
17
  },
18
  {
19
+ "epoch": 0.010526315789473684,
20
+ "eval_loss": 1.281607985496521,
21
+ "eval_runtime": 13.6651,
22
+ "eval_samples_per_second": 7.318,
23
+ "eval_steps_per_second": 1.464,
24
  "step": 1
25
  },
26
  {
27
+ "epoch": 0.021052631578947368,
28
+ "grad_norm": 1.099178671836853,
29
  "learning_rate": 4e-05,
30
+ "loss": 1.2265,
 
 
 
 
 
 
 
 
31
  "step": 2
32
  },
33
  {
34
+ "epoch": 0.031578947368421054,
35
+ "grad_norm": 0.8365464806556702,
36
  "learning_rate": 6e-05,
37
+ "loss": 1.2875,
38
  "step": 3
39
  },
40
  {
41
+ "epoch": 0.031578947368421054,
42
+ "eval_loss": 1.2764880657196045,
43
+ "eval_runtime": 13.7342,
44
+ "eval_samples_per_second": 7.281,
45
+ "eval_steps_per_second": 1.456,
46
+ "step": 3
47
  },
48
  {
49
+ "epoch": 0.042105263157894736,
50
+ "grad_norm": 0.7803714871406555,
51
+ "learning_rate": 8e-05,
52
+ "loss": 1.6142,
 
53
  "step": 4
54
  },
55
  {
56
+ "epoch": 0.05263157894736842,
57
+ "grad_norm": 0.7291099429130554,
58
  "learning_rate": 0.0001,
59
+ "loss": 1.2196,
60
  "step": 5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
61
  }
62
  ],
63
  "logging_steps": 1,
 
72
  "should_evaluate": false,
73
  "should_log": false,
74
  "should_save": true,
75
+ "should_training_stop": false
76
  },
77
  "attributes": {}
78
  }
79
  },
80
+ "total_flos": 2419306423910400.0,
81
+ "train_batch_size": 5,
82
  "trial_name": null,
83
  "trial_params": null
84
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d505e03c87584f88db69c68d4d3c92d4afd1945e9110366729a610299a3612d8
3
  size 6648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:166abb0c7e40600a59b0bdee3ffd387689a238ce61ab902f8e45f70ef598ee44
3
  size 6648