Kei5uke commited on
Commit
596a32c
·
verified ·
1 Parent(s): fa5605c

Training in progress, step 12, checkpoint

Browse files
checkpoint-12/adapter_config.json CHANGED
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "k_proj",
27
- "o_proj",
28
  "v_proj",
29
- "gate_proj",
 
30
  "up_proj",
31
- "down_proj",
32
- "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "down_proj",
 
27
  "v_proj",
28
+ "o_proj",
29
+ "q_proj",
30
  "up_proj",
31
+ "gate_proj",
32
+ "k_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
checkpoint-12/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c81ba4577f2b90ebf944c08cd92158beaf7c433a56828aeb2d2670b1ac57f2d2
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2efbc339582e5e7f609dd0db5aa19798900afe8df90e002a1e00dabb9fb3f77
3
  size 83945296
checkpoint-12/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc3066cc1d1029a905ce53416b23d2545c51cd844adc7f6327bfd833b2a1ae83
3
  size 168149074
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:037a71fde6d575f5374cea59eddced44e3a7c8271bdeaa43eaa2583f7c987f53
3
  size 168149074
checkpoint-12/trainer_state.json CHANGED
@@ -10,118 +10,118 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
- "grad_norm": 0.7731499075889587,
14
  "learning_rate": 4.000000000000001e-06,
15
- "loss": 1.166,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
- "grad_norm": 0.346652090549469,
21
  "learning_rate": 8.000000000000001e-06,
22
- "loss": 0.9594,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
- "grad_norm": 0.39937612414360046,
28
  "learning_rate": 1.2e-05,
29
- "loss": 0.828,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
- "eval_loss": 1.0491173267364502,
35
- "eval_runtime": 26.0459,
36
- "eval_samples_per_second": 3.839,
37
- "eval_steps_per_second": 3.839,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
- "grad_norm": 0.3457990288734436,
43
  "learning_rate": 1.6000000000000003e-05,
44
- "loss": 0.988,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
- "grad_norm": 0.513918399810791,
50
  "learning_rate": 2e-05,
51
- "loss": 0.8319,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
- "grad_norm": 0.47683727741241455,
57
  "learning_rate": 1.9200000000000003e-05,
58
- "loss": 0.9366,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
- "eval_loss": 1.0431987047195435,
64
- "eval_runtime": 26.2352,
65
- "eval_samples_per_second": 3.812,
66
- "eval_steps_per_second": 3.812,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
- "grad_norm": 0.4349649250507355,
72
  "learning_rate": 1.8400000000000003e-05,
73
- "loss": 0.9076,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
- "grad_norm": 0.5443339943885803,
79
  "learning_rate": 1.76e-05,
80
- "loss": 0.9829,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
- "grad_norm": 0.3990451395511627,
86
  "learning_rate": 1.6800000000000002e-05,
87
- "loss": 0.9045,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
- "eval_loss": 1.0292819738388062,
93
- "eval_runtime": 26.3477,
94
- "eval_samples_per_second": 3.795,
95
- "eval_steps_per_second": 3.795,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
- "grad_norm": 0.44750896096229553,
101
  "learning_rate": 1.6000000000000003e-05,
102
- "loss": 1.0768,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
- "grad_norm": 0.6865904331207275,
108
  "learning_rate": 1.5200000000000002e-05,
109
- "loss": 0.8717,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
- "grad_norm": 0.548491895198822,
115
  "learning_rate": 1.4400000000000001e-05,
116
- "loss": 1.1758,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
- "eval_loss": 1.0153621435165405,
122
- "eval_runtime": 26.5025,
123
- "eval_samples_per_second": 3.773,
124
- "eval_steps_per_second": 3.773,
125
  "step": 12
126
  }
127
  ],
@@ -142,7 +142,7 @@
142
  "attributes": {}
143
  }
144
  },
145
- "total_flos": 1.3299793854087168e+16,
146
  "train_batch_size": 2,
147
  "trial_name": null,
148
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0.0006200589055960316,
13
+ "grad_norm": 0.7607094645500183,
14
  "learning_rate": 4.000000000000001e-06,
15
+ "loss": 1.1893,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.0012401178111920632,
20
+ "grad_norm": 0.3518824279308319,
21
  "learning_rate": 8.000000000000001e-06,
22
+ "loss": 0.9752,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 0.0018601767167880949,
27
+ "grad_norm": 0.4024870693683624,
28
  "learning_rate": 1.2e-05,
29
+ "loss": 0.8467,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 0.0018601767167880949,
34
+ "eval_loss": 1.0757174491882324,
35
+ "eval_runtime": 26.414,
36
+ "eval_samples_per_second": 3.786,
37
+ "eval_steps_per_second": 3.786,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0024802356223841263,
42
+ "grad_norm": 0.36653462052345276,
43
  "learning_rate": 1.6000000000000003e-05,
44
+ "loss": 1.0056,
45
  "step": 4
46
  },
47
  {
48
  "epoch": 0.0031002945279801583,
49
+ "grad_norm": 0.47098514437675476,
50
  "learning_rate": 2e-05,
51
+ "loss": 0.8545,
52
  "step": 5
53
  },
54
  {
55
  "epoch": 0.0037203534335761897,
56
+ "grad_norm": 0.5030912160873413,
57
  "learning_rate": 1.9200000000000003e-05,
58
+ "loss": 0.9553,
59
  "step": 6
60
  },
61
  {
62
  "epoch": 0.0037203534335761897,
63
+ "eval_loss": 1.0662517547607422,
64
+ "eval_runtime": 26.7124,
65
+ "eval_samples_per_second": 3.744,
66
+ "eval_steps_per_second": 3.744,
67
  "step": 6
68
  },
69
  {
70
  "epoch": 0.004340412339172222,
71
+ "grad_norm": 0.4823940694332123,
72
  "learning_rate": 1.8400000000000003e-05,
73
+ "loss": 0.9266,
74
  "step": 7
75
  },
76
  {
77
  "epoch": 0.004960471244768253,
78
+ "grad_norm": 0.5647093653678894,
79
  "learning_rate": 1.76e-05,
80
+ "loss": 1.0013,
81
  "step": 8
82
  },
83
  {
84
  "epoch": 0.005580530150364285,
85
+ "grad_norm": 0.4058099687099457,
86
  "learning_rate": 1.6800000000000002e-05,
87
+ "loss": 0.9206,
88
  "step": 9
89
  },
90
  {
91
  "epoch": 0.005580530150364285,
92
+ "eval_loss": 1.0476710796356201,
93
+ "eval_runtime": 26.7708,
94
+ "eval_samples_per_second": 3.735,
95
+ "eval_steps_per_second": 3.735,
96
  "step": 9
97
  },
98
  {
99
  "epoch": 0.0062005890559603165,
100
+ "grad_norm": 0.44608208537101746,
101
  "learning_rate": 1.6000000000000003e-05,
102
+ "loss": 1.0937,
103
  "step": 10
104
  },
105
  {
106
  "epoch": 0.0068206479615563476,
107
+ "grad_norm": 0.7088199853897095,
108
  "learning_rate": 1.5200000000000002e-05,
109
+ "loss": 0.8915,
110
  "step": 11
111
  },
112
  {
113
  "epoch": 0.0074407068671523795,
114
+ "grad_norm": 0.5858705639839172,
115
  "learning_rate": 1.4400000000000001e-05,
116
+ "loss": 1.186,
117
  "step": 12
118
  },
119
  {
120
  "epoch": 0.0074407068671523795,
121
+ "eval_loss": 1.0301631689071655,
122
+ "eval_runtime": 27.1639,
123
+ "eval_samples_per_second": 3.681,
124
+ "eval_steps_per_second": 3.681,
125
  "step": 12
126
  }
127
  ],
 
142
  "attributes": {}
143
  }
144
  },
145
+ "total_flos": 1.3496219744698368e+16,
146
  "train_batch_size": 2,
147
  "trial_name": null,
148
  "trial_params": null
checkpoint-12/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cbe2fb31cf4c04f893e5299faa77267af76b8447f922d82c2d6c0200e1fe225f
3
  size 5688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:527a35da8257b0136b41661334383dc9a90d4f2271ec7e0b37f6fe024a71f4c9
3
  size 5688