Kei5uke commited on
Commit
85d14f7
·
verified ·
1 Parent(s): 00a3405

Training in progress, step 12, checkpoint

Browse files
checkpoint-12/adapter_config.json CHANGED
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "k_proj",
27
- "o_proj",
28
  "v_proj",
29
- "gate_proj",
 
30
  "up_proj",
31
- "down_proj",
32
- "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "down_proj",
 
27
  "v_proj",
28
+ "o_proj",
29
+ "q_proj",
30
  "up_proj",
31
+ "gate_proj",
32
+ "k_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
checkpoint-12/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56000b728927322d9e86b8406ce9fad531e6c2668058306b8dde96f73edbd533
3
  size 80013120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfdbb07538a47b1f579a81425d153481c6bf7e8b2b4b9e9312c4f6a93f54981a
3
  size 80013120
checkpoint-12/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1fec7b939776120d9a838393230d787fc8c6d4ef77a0ec38f3465a1b4efca68d
3
  size 160284754
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4f01ff1f2816fdfff56c7f358509b2bd9937f359e8ba3e2d168334d44dfa80d
3
  size 160284754
checkpoint-12/trainer_state.json CHANGED
@@ -10,118 +10,118 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
- "grad_norm": 0.18599237501621246,
14
  "learning_rate": 4.000000000000001e-06,
15
- "loss": 0.9319,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
- "grad_norm": 0.09086523950099945,
21
  "learning_rate": 8.000000000000001e-06,
22
- "loss": 0.6463,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
- "grad_norm": 0.09325356036424637,
28
  "learning_rate": 1.2e-05,
29
- "loss": 0.575,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
- "eval_loss": 0.7272388935089111,
35
- "eval_runtime": 33.9288,
36
- "eval_samples_per_second": 2.947,
37
- "eval_steps_per_second": 2.947,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
- "grad_norm": 0.09707680344581604,
43
  "learning_rate": 1.6000000000000003e-05,
44
- "loss": 0.7038,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
- "grad_norm": 0.10810782760381699,
50
  "learning_rate": 2e-05,
51
- "loss": 0.5471,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
- "grad_norm": 0.13030411303043365,
57
  "learning_rate": 1.9200000000000003e-05,
58
- "loss": 0.6735,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
- "eval_loss": 0.7256982922554016,
64
- "eval_runtime": 33.9528,
65
- "eval_samples_per_second": 2.945,
66
- "eval_steps_per_second": 2.945,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
- "grad_norm": 0.12336976826190948,
72
  "learning_rate": 1.8400000000000003e-05,
73
- "loss": 0.654,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
- "grad_norm": 0.16046489775180817,
79
  "learning_rate": 1.76e-05,
80
- "loss": 0.6301,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
- "grad_norm": 0.11589628458023071,
86
  "learning_rate": 1.6800000000000002e-05,
87
- "loss": 0.5326,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
- "eval_loss": 0.723076343536377,
93
- "eval_runtime": 34.5303,
94
- "eval_samples_per_second": 2.896,
95
- "eval_steps_per_second": 2.896,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
- "grad_norm": 0.12244553864002228,
101
  "learning_rate": 1.6000000000000003e-05,
102
- "loss": 0.8055,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
- "grad_norm": 0.10270956158638,
108
  "learning_rate": 1.5200000000000002e-05,
109
- "loss": 0.5492,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
- "grad_norm": 0.14439155161380768,
115
  "learning_rate": 1.4400000000000001e-05,
116
- "loss": 0.8579,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
- "eval_loss": 0.7205619215965271,
122
- "eval_runtime": 33.7297,
123
- "eval_samples_per_second": 2.965,
124
- "eval_steps_per_second": 2.965,
125
  "step": 12
126
  }
127
  ],
@@ -142,7 +142,7 @@
142
  "attributes": {}
143
  }
144
  },
145
- "total_flos": 1.5745185726971904e+16,
146
  "train_batch_size": 2,
147
  "trial_name": null,
148
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
+ "grad_norm": 0.16878783702850342,
14
  "learning_rate": 4.000000000000001e-06,
15
+ "loss": 0.9729,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
+ "grad_norm": 0.0865061804652214,
21
  "learning_rate": 8.000000000000001e-06,
22
+ "loss": 0.6661,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
+ "grad_norm": 0.0920981764793396,
28
  "learning_rate": 1.2e-05,
29
+ "loss": 0.5954,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
+ "eval_loss": 0.762992262840271,
35
+ "eval_runtime": 33.3972,
36
+ "eval_samples_per_second": 2.994,
37
+ "eval_steps_per_second": 2.994,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
+ "grad_norm": 0.10451442003250122,
43
  "learning_rate": 1.6000000000000003e-05,
44
+ "loss": 0.7227,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
+ "grad_norm": 0.11351602524518967,
50
  "learning_rate": 2e-05,
51
+ "loss": 0.5719,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
+ "grad_norm": 0.12484906613826752,
57
  "learning_rate": 1.9200000000000003e-05,
58
+ "loss": 0.6972,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
+ "eval_loss": 0.7615777850151062,
64
+ "eval_runtime": 33.3804,
65
+ "eval_samples_per_second": 2.996,
66
+ "eval_steps_per_second": 2.996,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
+ "grad_norm": 0.11116621643304825,
72
  "learning_rate": 1.8400000000000003e-05,
73
+ "loss": 0.6789,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
+ "grad_norm": 0.16013608872890472,
79
  "learning_rate": 1.76e-05,
80
+ "loss": 0.6612,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
+ "grad_norm": 0.1211152970790863,
86
  "learning_rate": 1.6800000000000002e-05,
87
+ "loss": 0.5557,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
+ "eval_loss": 0.7586225867271423,
93
+ "eval_runtime": 33.7015,
94
+ "eval_samples_per_second": 2.967,
95
+ "eval_steps_per_second": 2.967,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
+ "grad_norm": 0.1299518495798111,
101
  "learning_rate": 1.6000000000000003e-05,
102
+ "loss": 0.8339,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
+ "grad_norm": 0.11476285755634308,
108
  "learning_rate": 1.5200000000000002e-05,
109
+ "loss": 0.5786,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
+ "grad_norm": 0.18253028392791748,
115
  "learning_rate": 1.4400000000000001e-05,
116
+ "loss": 0.8889,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
+ "eval_loss": 0.7548640966415405,
122
+ "eval_runtime": 34.763,
123
+ "eval_samples_per_second": 2.877,
124
+ "eval_steps_per_second": 2.877,
125
  "step": 12
126
  }
127
  ],
 
142
  "attributes": {}
143
  }
144
  },
145
+ "total_flos": 1.5897880968413184e+16,
146
  "train_batch_size": 2,
147
  "trial_name": null,
148
  "trial_params": null
checkpoint-12/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd165a72088a81e5441087b0ae01dd72f7f14c5dad6a2e407eebc961586b0fc2
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fba20505ee5f941c5085404aed8036910f9919d2bc7b2ba5ba4186849388b0c
3
  size 5752