khoon485 commited on
Commit
07a9b76
·
1 Parent(s): 95b08fc
Files changed (7) hide show
  1. config.json +1 -1
  2. optimizer.pt +1 -1
  3. pytorch_model.bin +1 -1
  4. rng_state.pth +1 -1
  5. scheduler.pt +1 -1
  6. trainer_state.json +69 -99
  7. training_args.bin +1 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "t5-base",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
 
1
  {
2
+ "_name_or_path": "/home/ubuntu/Pythonprojects/kcbert2/models/t5/t5-legacy/",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e7250ce1ab70dc1a9fd526bb4d5d8eb5222e08cdf5b654e319428c72c219d89
3
  size 1783379133
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dda6fb47598a3c4d7e0183f11bcc07d7b518fcb40f9facd62984206e748d3a01
3
  size 1783379133
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e23f9bb16399dbb6299333937b7d2eeb28058c49c33ee3ff3a2e4f28f9930cb
3
  size 891730879
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:427d75649726c7fc41848295a130297c7ec168feb884b0560d83dfdda7cfb76a
3
  size 891730879
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13d5b1bae4964fb2f4d01538a124e8fa3d09ef60275a74b9b04e9572c2256028
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b95803f7c7708b5b515d43cad3bc8ce32a703e43dd162abb60ddc0b762fca184
3
  size 14503
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06cbc2ad564cf4cd4ae1683a2a75297d61e4bb88322c20b1f9c2c3540e65749c
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccc836a9c4b5ccd266a159052e75698e88fc2f725341bb9a62ac20a2edf57466
3
  size 623
trainer_state.json CHANGED
@@ -1,146 +1,116 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 14.838626976214215,
5
- "global_step": 6500,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
  "epoch": 1.0,
12
- "eval_loss": 0.9184420704841614,
13
- "eval_runtime": 226.0895,
14
- "eval_samples_per_second": 110.098,
15
- "eval_steps_per_second": 27.524,
16
- "step": 438
17
  },
18
  {
19
  "epoch": 2.0,
20
- "eval_loss": 0.8269533514976501,
21
- "eval_runtime": 222.6047,
22
- "eval_samples_per_second": 111.822,
23
- "eval_steps_per_second": 27.955,
24
- "step": 876
25
  },
26
  {
27
  "epoch": 3.0,
28
- "eval_loss": 0.7799285054206848,
29
- "eval_runtime": 267.0428,
30
- "eval_samples_per_second": 93.214,
31
- "eval_steps_per_second": 23.303,
32
- "step": 1314
33
  },
34
  {
35
  "epoch": 4.0,
36
- "eval_loss": 0.7528017163276672,
37
- "eval_runtime": 225.2214,
38
- "eval_samples_per_second": 110.522,
39
- "eval_steps_per_second": 27.631,
40
- "step": 1752
41
  },
42
  {
43
- "epoch": 4.57,
44
- "learning_rate": 6.9558599695586e-05,
45
- "loss": 0.9432,
46
- "step": 2000
 
 
47
  },
48
  {
49
- "epoch": 5.0,
50
- "eval_loss": 0.7184237241744995,
51
- "eval_runtime": 225.9905,
52
- "eval_samples_per_second": 110.146,
53
- "eval_steps_per_second": 27.537,
54
- "step": 2190
55
  },
56
  {
57
  "epoch": 6.0,
58
- "eval_loss": 0.6917603611946106,
59
- "eval_runtime": 229.2094,
60
- "eval_samples_per_second": 108.599,
61
- "eval_steps_per_second": 27.15,
62
- "step": 2628
63
  },
64
  {
65
  "epoch": 7.0,
66
- "eval_loss": 0.6674416065216064,
67
- "eval_runtime": 228.8661,
68
- "eval_samples_per_second": 108.762,
69
- "eval_steps_per_second": 27.191,
70
- "step": 3066
71
  },
72
  {
73
  "epoch": 8.0,
74
- "eval_loss": 0.6534304022789001,
75
- "eval_runtime": 238.4381,
76
- "eval_samples_per_second": 104.396,
77
- "eval_steps_per_second": 26.099,
78
- "step": 3504
79
  },
80
  {
81
  "epoch": 9.0,
82
- "eval_loss": 0.6406867504119873,
83
- "eval_runtime": 235.4552,
84
- "eval_samples_per_second": 105.719,
85
- "eval_steps_per_second": 26.43,
86
- "step": 3942
87
- },
88
- {
89
- "epoch": 9.13,
90
- "learning_rate": 3.9117199391171995e-05,
91
- "loss": 0.6607,
92
- "step": 4000
93
  },
94
  {
95
  "epoch": 10.0,
96
- "eval_loss": 0.6275271773338318,
97
- "eval_runtime": 231.6414,
98
- "eval_samples_per_second": 107.459,
99
- "eval_steps_per_second": 26.865,
100
- "step": 4380
101
  },
102
  {
103
  "epoch": 11.0,
104
- "eval_loss": 0.6164868474006653,
105
- "eval_runtime": 229.8159,
106
- "eval_samples_per_second": 108.313,
107
- "eval_steps_per_second": 27.078,
108
- "step": 4818
109
- },
110
- {
111
- "epoch": 12.0,
112
- "eval_loss": 0.610621452331543,
113
- "eval_runtime": 234.5136,
114
- "eval_samples_per_second": 106.143,
115
- "eval_steps_per_second": 26.536,
116
- "step": 5256
117
- },
118
- {
119
- "epoch": 13.0,
120
- "eval_loss": 0.6076993942260742,
121
- "eval_runtime": 221.8665,
122
- "eval_samples_per_second": 112.194,
123
- "eval_steps_per_second": 28.048,
124
- "step": 5694
125
- },
126
- {
127
- "epoch": 13.7,
128
- "learning_rate": 8.675799086757991e-06,
129
- "loss": 0.5624,
130
- "step": 6000
131
  },
132
  {
133
- "epoch": 14.0,
134
- "eval_loss": 0.6030386686325073,
135
- "eval_runtime": 225.3038,
136
- "eval_samples_per_second": 110.482,
137
- "eval_steps_per_second": 27.62,
138
- "step": 6132
139
  }
140
  ],
141
- "max_steps": 6570,
142
- "num_train_epochs": 15,
143
- "total_flos": 5.075204254610227e+17,
144
  "trial_name": null,
145
  "trial_params": null
146
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 11.965708791795748,
5
+ "global_step": 21000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
  "epoch": 1.0,
12
+ "eval_loss": 0.5735206007957458,
13
+ "eval_runtime": 226.9724,
14
+ "eval_samples_per_second": 109.67,
15
+ "eval_steps_per_second": 27.417,
16
+ "step": 1755
17
  },
18
  {
19
  "epoch": 2.0,
20
+ "eval_loss": 0.5326883792877197,
21
+ "eval_runtime": 225.773,
22
+ "eval_samples_per_second": 110.252,
23
+ "eval_steps_per_second": 27.563,
24
+ "step": 3510
25
  },
26
  {
27
  "epoch": 3.0,
28
+ "eval_loss": 0.5004657506942749,
29
+ "eval_runtime": 223.9431,
30
+ "eval_samples_per_second": 111.153,
31
+ "eval_steps_per_second": 27.788,
32
+ "step": 5265
33
  },
34
  {
35
  "epoch": 4.0,
36
+ "eval_loss": 0.48436784744262695,
37
+ "eval_runtime": 223.6051,
38
+ "eval_samples_per_second": 111.321,
39
+ "eval_steps_per_second": 27.83,
40
+ "step": 7020
41
  },
42
  {
43
+ "epoch": 5.0,
44
+ "eval_loss": 0.4659739136695862,
45
+ "eval_runtime": 221.9535,
46
+ "eval_samples_per_second": 112.15,
47
+ "eval_steps_per_second": 28.037,
48
+ "step": 8775
49
  },
50
  {
51
+ "epoch": 5.7,
52
+ "learning_rate": 5.251661918328585e-05,
53
+ "loss": 0.4274,
54
+ "step": 10000
 
 
55
  },
56
  {
57
  "epoch": 6.0,
58
+ "eval_loss": 0.4595305621623993,
59
+ "eval_runtime": 232.1017,
60
+ "eval_samples_per_second": 107.246,
61
+ "eval_steps_per_second": 26.812,
62
+ "step": 10530
63
  },
64
  {
65
  "epoch": 7.0,
66
+ "eval_loss": 0.44809117913246155,
67
+ "eval_runtime": 230.3824,
68
+ "eval_samples_per_second": 108.046,
69
+ "eval_steps_per_second": 27.012,
70
+ "step": 12285
71
  },
72
  {
73
  "epoch": 8.0,
74
+ "eval_loss": 0.44372686743736267,
75
+ "eval_runtime": 220.8242,
76
+ "eval_samples_per_second": 112.723,
77
+ "eval_steps_per_second": 28.181,
78
+ "step": 14040
79
  },
80
  {
81
  "epoch": 9.0,
82
+ "eval_loss": 0.4426543712615967,
83
+ "eval_runtime": 226.8705,
84
+ "eval_samples_per_second": 109.719,
85
+ "eval_steps_per_second": 27.43,
86
+ "step": 15795
 
 
 
 
 
 
87
  },
88
  {
89
  "epoch": 10.0,
90
+ "eval_loss": 0.4342004656791687,
91
+ "eval_runtime": 221.9982,
92
+ "eval_samples_per_second": 112.127,
93
+ "eval_steps_per_second": 28.032,
94
+ "step": 17550
95
  },
96
  {
97
  "epoch": 11.0,
98
+ "eval_loss": 0.43299490213394165,
99
+ "eval_runtime": 221.9453,
100
+ "eval_samples_per_second": 112.154,
101
+ "eval_steps_per_second": 28.038,
102
+ "step": 19305
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
103
  },
104
  {
105
+ "epoch": 11.4,
106
+ "learning_rate": 5.0332383665717e-06,
107
+ "loss": 0.2777,
108
+ "step": 20000
 
 
109
  }
110
  ],
111
+ "max_steps": 21060,
112
+ "num_train_epochs": 12,
113
+ "total_flos": 4.092582190428979e+17,
114
  "trial_name": null,
115
  "trial_params": null
116
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92440562452781ca3f73efd16bc795762158e4972a2597602a92148285739f94
3
  size 2991
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5570b0e849e910b9f449b5cd2078e8f63540a7824db3dbf9415c2f63993b48b3
3
  size 2991