handeboyaci commited on
Commit
f0b182f
·
verified ·
1 Parent(s): efd50ef

Upload folder using huggingface_hub

Browse files
Files changed (7) hide show
  1. config.json +57 -0
  2. model.safetensors +3 -0
  3. optimizer.pt +3 -0
  4. rng_state.pth +3 -0
  5. scheduler.pt +3 -0
  6. trainer_state.json +196 -0
  7. training_args.bin +3 -0
config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "clefourrier/graphormer-base-pcqm4mv2",
3
+ "activation_dropout": 0.1,
4
+ "activation_fn": "gelu",
5
+ "apply_graphormer_init": true,
6
+ "architectures": [
7
+ "GraphormerForGraphClassification"
8
+ ],
9
+ "attention_dropout": 0.1,
10
+ "bias": true,
11
+ "bos_token_id": 1,
12
+ "dropout": 0.0,
13
+ "edge_type": "multi_hop",
14
+ "embed_scale": null,
15
+ "embedding_dim": 768,
16
+ "encoder_normalize_before": true,
17
+ "eos_token_id": 2,
18
+ "export": false,
19
+ "ffn_embedding_dim": 768,
20
+ "freeze_embeddings": false,
21
+ "hidden_size": 768,
22
+ "id2label": {
23
+ "0": "LABEL_0"
24
+ },
25
+ "init_fn": null,
26
+ "kdim": null,
27
+ "label2id": {
28
+ "LABEL_0": 0
29
+ },
30
+ "layerdrop": 0.0,
31
+ "max_nodes": 512,
32
+ "model_type": "graphormer",
33
+ "multi_hop_max_dist": 5,
34
+ "no_token_positional_embeddings": false,
35
+ "num_atoms": 4608,
36
+ "num_attention_heads": 32,
37
+ "num_classes": 2,
38
+ "num_edge_dis": 128,
39
+ "num_edges": 1536,
40
+ "num_hidden_layers": 12,
41
+ "num_in_degree": 512,
42
+ "num_layers": 12,
43
+ "num_out_degree": 512,
44
+ "num_spatial": 512,
45
+ "num_trans_layers_to_freeze": 0,
46
+ "pad_token_id": 0,
47
+ "pre_layernorm": false,
48
+ "q_noise": 0.0,
49
+ "qn_block_size": 8,
50
+ "self_attention": true,
51
+ "share_input_output_embed": false,
52
+ "spatial_pos_max": 1024,
53
+ "torch_dtype": "float32",
54
+ "traceable": false,
55
+ "transformers_version": "4.40.2",
56
+ "vdim": null
57
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48ecfae273235162997a20c2aefce41c99b18fd5c4a0a6d94a49a53ea79ca13e
3
+ size 190738260
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6de090cabf90665821d775934b39b6e11e035838ec861ff5bbdfaa2e5de5653
3
+ size 381600651
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fee6d93b2488d942c7e32572a01fbe632a7d98db8db4d0e34443c0f58a3f7f33
3
+ size 14645
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e73fe0c8fb778eafc7f155cc1be97d53fe3d6292a71cfeba42689e6ce858e6f3
3
+ size 1465
trainer_state.json ADDED
@@ -0,0 +1,196 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8207028311058135,
3
+ "best_model_checkpoint": "/content/drive/MyDrive/BBBP_results/graphormer/checkpoint-686",
4
+ "epoch": 7.0,
5
+ "eval_steps": 500,
6
+ "global_step": 686,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.5102040816326531,
13
+ "grad_norm": 5.745197296142578,
14
+ "learning_rate": 1.8979591836734696e-05,
15
+ "loss": 0.5765,
16
+ "step": 50
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_accuracy": 0.6702944942381562,
21
+ "eval_f1": 0.7691618108471537,
22
+ "eval_loss": 0.7216734886169434,
23
+ "eval_mcc": 0.3080880924518934,
24
+ "eval_roc_auc": 0.7584252660766063,
25
+ "eval_runtime": 174.1926,
26
+ "eval_samples_per_second": 8.967,
27
+ "eval_steps_per_second": 0.144,
28
+ "step": 98
29
+ },
30
+ {
31
+ "epoch": 1.0204081632653061,
32
+ "grad_norm": 4.329917907714844,
33
+ "learning_rate": 1.795918367346939e-05,
34
+ "loss": 0.395,
35
+ "step": 100
36
+ },
37
+ {
38
+ "epoch": 1.5306122448979593,
39
+ "grad_norm": 7.502938747406006,
40
+ "learning_rate": 1.6938775510204085e-05,
41
+ "loss": 0.3451,
42
+ "step": 150
43
+ },
44
+ {
45
+ "epoch": 2.0,
46
+ "eval_accuracy": 0.7144686299615877,
47
+ "eval_f1": 0.7747474747474747,
48
+ "eval_loss": 0.6415975093841553,
49
+ "eval_mcc": 0.3986338998981387,
50
+ "eval_roc_auc": 0.7898015943462421,
51
+ "eval_runtime": 166.1136,
52
+ "eval_samples_per_second": 9.403,
53
+ "eval_steps_per_second": 0.15,
54
+ "step": 196
55
+ },
56
+ {
57
+ "epoch": 2.0408163265306123,
58
+ "grad_norm": 5.854671001434326,
59
+ "learning_rate": 1.5918367346938776e-05,
60
+ "loss": 0.3423,
61
+ "step": 200
62
+ },
63
+ {
64
+ "epoch": 2.5510204081632653,
65
+ "grad_norm": 6.938652038574219,
66
+ "learning_rate": 1.4897959183673472e-05,
67
+ "loss": 0.3229,
68
+ "step": 250
69
+ },
70
+ {
71
+ "epoch": 3.0,
72
+ "eval_accuracy": 0.7400768245838668,
73
+ "eval_f1": 0.7883211678832117,
74
+ "eval_loss": 0.5721713304519653,
75
+ "eval_mcc": 0.4562663481861473,
76
+ "eval_roc_auc": 0.8044279881311658,
77
+ "eval_runtime": 165.7968,
78
+ "eval_samples_per_second": 9.421,
79
+ "eval_steps_per_second": 0.151,
80
+ "step": 294
81
+ },
82
+ {
83
+ "epoch": 3.061224489795918,
84
+ "grad_norm": 4.135607719421387,
85
+ "learning_rate": 1.3877551020408165e-05,
86
+ "loss": 0.3051,
87
+ "step": 300
88
+ },
89
+ {
90
+ "epoch": 3.571428571428571,
91
+ "grad_norm": 12.983510971069336,
92
+ "learning_rate": 1.2857142857142859e-05,
93
+ "loss": 0.2861,
94
+ "step": 350
95
+ },
96
+ {
97
+ "epoch": 4.0,
98
+ "eval_accuracy": 0.7451984635083226,
99
+ "eval_f1": 0.7876200640341515,
100
+ "eval_loss": 0.5572524070739746,
101
+ "eval_mcc": 0.4702510049659752,
102
+ "eval_roc_auc": 0.808828924788449,
103
+ "eval_runtime": 165.8102,
104
+ "eval_samples_per_second": 9.42,
105
+ "eval_steps_per_second": 0.151,
106
+ "step": 392
107
+ },
108
+ {
109
+ "epoch": 4.081632653061225,
110
+ "grad_norm": 4.853840351104736,
111
+ "learning_rate": 1.1836734693877552e-05,
112
+ "loss": 0.2916,
113
+ "step": 400
114
+ },
115
+ {
116
+ "epoch": 4.591836734693878,
117
+ "grad_norm": 12.244647026062012,
118
+ "learning_rate": 1.0816326530612246e-05,
119
+ "loss": 0.2669,
120
+ "step": 450
121
+ },
122
+ {
123
+ "epoch": 5.0,
124
+ "eval_accuracy": 0.7298335467349552,
125
+ "eval_f1": 0.7923228346456693,
126
+ "eval_loss": 0.585197389125824,
127
+ "eval_mcc": 0.432741923477754,
128
+ "eval_roc_auc": 0.811532381459596,
129
+ "eval_runtime": 165.2833,
130
+ "eval_samples_per_second": 9.45,
131
+ "eval_steps_per_second": 0.151,
132
+ "step": 490
133
+ },
134
+ {
135
+ "epoch": 5.1020408163265305,
136
+ "grad_norm": 6.616086959838867,
137
+ "learning_rate": 9.795918367346939e-06,
138
+ "loss": 0.2686,
139
+ "step": 500
140
+ },
141
+ {
142
+ "epoch": 5.612244897959184,
143
+ "grad_norm": 4.586119174957275,
144
+ "learning_rate": 8.775510204081633e-06,
145
+ "loss": 0.2575,
146
+ "step": 550
147
+ },
148
+ {
149
+ "epoch": 6.0,
150
+ "eval_accuracy": 0.7215108834827144,
151
+ "eval_f1": 0.7895500725689405,
152
+ "eval_loss": 0.6616533994674683,
153
+ "eval_mcc": 0.4156650698011332,
154
+ "eval_roc_auc": 0.8114326291496539,
155
+ "eval_runtime": 165.0772,
156
+ "eval_samples_per_second": 9.462,
157
+ "eval_steps_per_second": 0.151,
158
+ "step": 588
159
+ },
160
+ {
161
+ "epoch": 6.122448979591836,
162
+ "grad_norm": 4.342059135437012,
163
+ "learning_rate": 7.755102040816327e-06,
164
+ "loss": 0.2691,
165
+ "step": 600
166
+ },
167
+ {
168
+ "epoch": 6.63265306122449,
169
+ "grad_norm": 8.41532039642334,
170
+ "learning_rate": 6.734693877551021e-06,
171
+ "loss": 0.2464,
172
+ "step": 650
173
+ },
174
+ {
175
+ "epoch": 7.0,
176
+ "eval_accuracy": 0.7573623559539052,
177
+ "eval_f1": 0.802088772845953,
178
+ "eval_loss": 0.5653724074363708,
179
+ "eval_mcc": 0.4929882749662992,
180
+ "eval_roc_auc": 0.8207028311058135,
181
+ "eval_runtime": 165.1778,
182
+ "eval_samples_per_second": 9.456,
183
+ "eval_steps_per_second": 0.151,
184
+ "step": 686
185
+ }
186
+ ],
187
+ "logging_steps": 50,
188
+ "max_steps": 980,
189
+ "num_input_tokens_seen": 0,
190
+ "num_train_epochs": 10,
191
+ "save_steps": 500,
192
+ "total_flos": 0.0,
193
+ "train_batch_size": 64,
194
+ "trial_name": null,
195
+ "trial_params": null
196
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85218b53ef2e13204c5b773ee6ee0def4cf0cc294c9e0b7b2bbed3839ddcb76a
3
+ size 5457