atomwalk12 commited on
Commit
d5bec65
·
verified ·
1 Parent(s): 8e5e4cb

End of training

Browse files
README.md CHANGED
@@ -5,9 +5,9 @@ library_name: transformers
5
  model_name: trainer_output
6
  tags:
7
  - generated_from_trainer
 
8
  - sft
9
  - unsloth
10
- - tool-use
11
  - linalg-zero
12
  - trl
13
  licence: license
 
5
  model_name: trainer_output
6
  tags:
7
  - generated_from_trainer
8
+ - tool-use
9
  - sft
10
  - unsloth
 
11
  - linalg-zero
12
  - trl
13
  licence: license
adapter_config.json CHANGED
@@ -29,13 +29,13 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
32
  "o_proj",
33
- "q_proj",
34
- "v_proj",
35
  "down_proj",
36
- "up_proj",
37
  "k_proj",
38
- "gate_proj"
 
 
39
  ],
40
  "task_type": "CAUSAL_LM",
41
  "trainable_token_indices": null,
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "gate_proj",
33
  "o_proj",
 
 
34
  "down_proj",
 
35
  "k_proj",
36
+ "q_proj",
37
+ "up_proj",
38
+ "v_proj"
39
  ],
40
  "task_type": "CAUSAL_LM",
41
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c0d12cda975b3ad62a4d859cb76757742b4ac86b08fa133763b92b51e0ffb5b6
3
  size 132188392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0156b8e2bd7ee00762484b1b56e008189907477eea7e35d71fabc479be41072
3
  size 132188392
all_results.json CHANGED
@@ -1,9 +1,14 @@
1
  {
2
  "epoch": 1.0,
 
 
 
 
 
3
  "total_flos": 2717720554039296.0,
4
- "train_loss": 0.24179711043834687,
5
- "train_runtime": 32.7348,
6
  "train_samples": 60,
7
- "train_samples_per_second": 1.833,
8
- "train_steps_per_second": 1.833
9
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 0.5140631794929504,
4
+ "eval_runtime": 8.3035,
5
+ "eval_samples": 60,
6
+ "eval_samples_per_second": 7.226,
7
+ "eval_steps_per_second": 7.226,
8
  "total_flos": 2717720554039296.0,
9
+ "train_loss": 0.24174172580242156,
10
+ "train_runtime": 31.6344,
11
  "train_samples": 60,
12
+ "train_samples_per_second": 1.897,
13
+ "train_steps_per_second": 1.897
14
  }
eval_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_loss": 0.5140631794929504,
4
+ "eval_runtime": 8.3035,
5
+ "eval_samples": 60,
6
+ "eval_samples_per_second": 7.226,
7
+ "eval_steps_per_second": 7.226
8
+ }
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 1.0,
3
  "total_flos": 2717720554039296.0,
4
- "train_loss": 0.24179711043834687,
5
- "train_runtime": 32.7348,
6
  "train_samples": 60,
7
- "train_samples_per_second": 1.833,
8
- "train_steps_per_second": 1.833
9
  }
 
1
  {
2
  "epoch": 1.0,
3
  "total_flos": 2717720554039296.0,
4
+ "train_loss": 0.24174172580242156,
5
+ "train_runtime": 31.6344,
6
  "train_samples": 60,
7
+ "train_samples_per_second": 1.897,
8
+ "train_steps_per_second": 1.897
9
  }
trainer_state.json CHANGED
@@ -11,96 +11,96 @@
11
  "log_history": [
12
  {
13
  "epoch": 0.08333333333333333,
14
- "grad_norm": 0.3455909192562103,
15
  "learning_rate": 0.00016,
16
- "loss": 0.7527,
17
  "step": 5
18
  },
19
  {
20
  "epoch": 0.16666666666666666,
21
- "grad_norm": 0.532439649105072,
22
  "learning_rate": 0.00018545454545454545,
23
- "loss": 0.426,
24
  "step": 10
25
  },
26
  {
27
  "epoch": 0.25,
28
- "grad_norm": 0.3362359404563904,
29
  "learning_rate": 0.00016727272727272728,
30
- "loss": 0.2268,
31
  "step": 15
32
  },
33
  {
34
  "epoch": 0.3333333333333333,
35
- "grad_norm": 0.2568519115447998,
36
  "learning_rate": 0.0001490909090909091,
37
- "loss": 0.1894,
38
  "step": 20
39
  },
40
  {
41
  "epoch": 0.4166666666666667,
42
- "grad_norm": 0.263459712266922,
43
  "learning_rate": 0.00013090909090909093,
44
- "loss": 0.1829,
45
  "step": 25
46
  },
47
  {
48
  "epoch": 0.5,
49
- "grad_norm": 0.20919214189052582,
50
  "learning_rate": 0.00011272727272727272,
51
- "loss": 0.1623,
52
  "step": 30
53
  },
54
  {
55
  "epoch": 0.5833333333333334,
56
- "grad_norm": 0.2152024358510971,
57
  "learning_rate": 9.454545454545455e-05,
58
  "loss": 0.1638,
59
  "step": 35
60
  },
61
  {
62
  "epoch": 0.6666666666666666,
63
- "grad_norm": 0.2593837082386017,
64
  "learning_rate": 7.636363636363637e-05,
65
- "loss": 0.1665,
66
  "step": 40
67
  },
68
  {
69
  "epoch": 0.75,
70
- "grad_norm": 0.20896115899085999,
71
  "learning_rate": 5.818181818181818e-05,
72
  "loss": 0.1678,
73
  "step": 45
74
  },
75
  {
76
  "epoch": 0.8333333333333334,
77
- "grad_norm": 0.2554057538509369,
78
  "learning_rate": 4e-05,
79
- "loss": 0.155,
80
  "step": 50
81
  },
82
  {
83
  "epoch": 0.9166666666666666,
84
- "grad_norm": 0.17573142051696777,
85
  "learning_rate": 2.1818181818181818e-05,
86
- "loss": 0.1529,
87
  "step": 55
88
  },
89
  {
90
  "epoch": 1.0,
91
- "grad_norm": 0.24647420644760132,
92
  "learning_rate": 3.636363636363636e-06,
93
- "loss": 0.1557,
94
  "step": 60
95
  },
96
  {
97
  "epoch": 1.0,
98
  "step": 60,
99
  "total_flos": 2717720554039296.0,
100
- "train_loss": 0.24179711043834687,
101
- "train_runtime": 32.7348,
102
- "train_samples_per_second": 1.833,
103
- "train_steps_per_second": 1.833
104
  }
105
  ],
106
  "logging_steps": 5,
 
11
  "log_history": [
12
  {
13
  "epoch": 0.08333333333333333,
14
+ "grad_norm": 0.3362419009208679,
15
  "learning_rate": 0.00016,
16
+ "loss": 0.7526,
17
  "step": 5
18
  },
19
  {
20
  "epoch": 0.16666666666666666,
21
+ "grad_norm": 0.5312803983688354,
22
  "learning_rate": 0.00018545454545454545,
23
+ "loss": 0.4235,
24
  "step": 10
25
  },
26
  {
27
  "epoch": 0.25,
28
+ "grad_norm": 0.3434021770954132,
29
  "learning_rate": 0.00016727272727272728,
30
+ "loss": 0.2288,
31
  "step": 15
32
  },
33
  {
34
  "epoch": 0.3333333333333333,
35
+ "grad_norm": 0.26192706823349,
36
  "learning_rate": 0.0001490909090909091,
37
+ "loss": 0.1897,
38
  "step": 20
39
  },
40
  {
41
  "epoch": 0.4166666666666667,
42
+ "grad_norm": 0.26772984862327576,
43
  "learning_rate": 0.00013090909090909093,
44
+ "loss": 0.1833,
45
  "step": 25
46
  },
47
  {
48
  "epoch": 0.5,
49
+ "grad_norm": 0.20748060941696167,
50
  "learning_rate": 0.00011272727272727272,
51
+ "loss": 0.1621,
52
  "step": 30
53
  },
54
  {
55
  "epoch": 0.5833333333333334,
56
+ "grad_norm": 0.21864831447601318,
57
  "learning_rate": 9.454545454545455e-05,
58
  "loss": 0.1638,
59
  "step": 35
60
  },
61
  {
62
  "epoch": 0.6666666666666666,
63
+ "grad_norm": 0.24830199778079987,
64
  "learning_rate": 7.636363636363637e-05,
65
+ "loss": 0.1667,
66
  "step": 40
67
  },
68
  {
69
  "epoch": 0.75,
70
+ "grad_norm": 0.20934940874576569,
71
  "learning_rate": 5.818181818181818e-05,
72
  "loss": 0.1678,
73
  "step": 45
74
  },
75
  {
76
  "epoch": 0.8333333333333334,
77
+ "grad_norm": 0.2636692523956299,
78
  "learning_rate": 4e-05,
79
+ "loss": 0.1548,
80
  "step": 50
81
  },
82
  {
83
  "epoch": 0.9166666666666666,
84
+ "grad_norm": 0.1722150295972824,
85
  "learning_rate": 2.1818181818181818e-05,
86
+ "loss": 0.1524,
87
  "step": 55
88
  },
89
  {
90
  "epoch": 1.0,
91
+ "grad_norm": 0.23989179730415344,
92
  "learning_rate": 3.636363636363636e-06,
93
+ "loss": 0.1553,
94
  "step": 60
95
  },
96
  {
97
  "epoch": 1.0,
98
  "step": 60,
99
  "total_flos": 2717720554039296.0,
100
+ "train_loss": 0.24174172580242156,
101
+ "train_runtime": 31.6344,
102
+ "train_samples_per_second": 1.897,
103
+ "train_steps_per_second": 1.897
104
  }
105
  ],
106
  "logging_steps": 5,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8e28e6b8f64da6914dafb8a55f4cfa8471104078dcab1dde68ecb85d89f925bb
3
  size 6097
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54b494cceb94d5932b6c823a353a9515458fd050c35e966606a622ff2776e061
3
  size 6097