msb-roshan kmchiti commited on
Commit
f1426a5
·
verified ·
0 Parent(s):

Duplicate from MolGen/llama_ZINC_1B-raw_atomwise_SAFE_a0df739a

Browse files

Co-authored-by: Kamran Chitsaz <kmchiti@users.noreply.huggingface.co>

This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +105 -0
  2. run_history.csv +0 -0
  3. tmp-spec-checkpoint-10000/config.json +30 -0
  4. tmp-spec-checkpoint-10000/generated_smiles.csv +0 -0
  5. tmp-spec-checkpoint-10000/generated_smiles_100k.csv +0 -0
  6. tmp-spec-checkpoint-10000/generated_smiles_100k_0.csv +0 -0
  7. tmp-spec-checkpoint-10000/generated_smiles_100k_1.csv +0 -0
  8. tmp-spec-checkpoint-10000/generated_smiles_100k_2.csv +0 -0
  9. tmp-spec-checkpoint-10000/generated_smiles_1M.csv +3 -0
  10. tmp-spec-checkpoint-10000/generated_smiles_1M_0.csv +3 -0
  11. tmp-spec-checkpoint-10000/generated_smiles_1M_1.csv +3 -0
  12. tmp-spec-checkpoint-10000/generated_smiles_1M_2.csv +3 -0
  13. tmp-spec-checkpoint-10000/generated_smiles_30k.csv +0 -0
  14. tmp-spec-checkpoint-10000/generated_smiles_30k_0.csv +0 -0
  15. tmp-spec-checkpoint-10000/generated_smiles_30k_1.csv +0 -0
  16. tmp-spec-checkpoint-10000/generated_smiles_30k_2.csv +0 -0
  17. tmp-spec-checkpoint-10000/pytorch_model.bin +3 -0
  18. tmp-spec-checkpoint-10000/result_FCD_SNN_Frag_Scaf.json +19 -0
  19. tmp-spec-checkpoint-10000/result_SA_wasserstein.json +7 -0
  20. tmp-spec-checkpoint-10000/result_novelty_total.json +3 -0
  21. tmp-spec-checkpoint-10000/result_unique@1k_unique@10k_IntDiv_IntDiv2_logP_wasserstein_QED_wasserstein_weight_wasserstein_NP_wasserstein.json +23 -0
  22. tmp-spec-checkpoint-10000/special_tokens_map.json +30 -0
  23. tmp-spec-checkpoint-10000/tokenizer.json +193 -0
  24. tmp-spec-checkpoint-10000/tokenizer_config.json +43 -0
  25. tmp-spec-checkpoint-10000/training_args.bin +3 -0
  26. tmp-spec-checkpoint-15000/config.json +30 -0
  27. tmp-spec-checkpoint-15000/generated_smiles.csv +0 -0
  28. tmp-spec-checkpoint-15000/generated_smiles_100k.csv +0 -0
  29. tmp-spec-checkpoint-15000/generated_smiles_100k_0.csv +0 -0
  30. tmp-spec-checkpoint-15000/generated_smiles_100k_1.csv +0 -0
  31. tmp-spec-checkpoint-15000/generated_smiles_100k_2.csv +0 -0
  32. tmp-spec-checkpoint-15000/generated_smiles_1M.csv +3 -0
  33. tmp-spec-checkpoint-15000/generated_smiles_1M_0.csv +3 -0
  34. tmp-spec-checkpoint-15000/generated_smiles_1M_1.csv +3 -0
  35. tmp-spec-checkpoint-15000/generated_smiles_1M_2.csv +3 -0
  36. tmp-spec-checkpoint-15000/generated_smiles_30k.csv +0 -0
  37. tmp-spec-checkpoint-15000/generated_smiles_30k_0.csv +0 -0
  38. tmp-spec-checkpoint-15000/generated_smiles_30k_1.csv +0 -0
  39. tmp-spec-checkpoint-15000/generated_smiles_30k_2.csv +0 -0
  40. tmp-spec-checkpoint-15000/pytorch_model.bin +3 -0
  41. tmp-spec-checkpoint-15000/result_FCD_SNN_Frag_Scaf.json +19 -0
  42. tmp-spec-checkpoint-15000/result_SA_wasserstein.json +7 -0
  43. tmp-spec-checkpoint-15000/result_novelty_total.json +3 -0
  44. tmp-spec-checkpoint-15000/result_unique@1k_unique@10k_IntDiv_IntDiv2_logP_wasserstein_QED_wasserstein_weight_wasserstein_NP_wasserstein.json +23 -0
  45. tmp-spec-checkpoint-15000/special_tokens_map.json +30 -0
  46. tmp-spec-checkpoint-15000/tokenizer.json +193 -0
  47. tmp-spec-checkpoint-15000/tokenizer_config.json +43 -0
  48. tmp-spec-checkpoint-15000/training_args.bin +3 -0
  49. tmp-spec-checkpoint-20000/config.json +30 -0
  50. tmp-spec-checkpoint-20000/generated_smiles.csv +0 -0
.gitattributes ADDED
@@ -0,0 +1,105 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tmp-spec-checkpoint-5000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
37
+ tmp-spec-checkpoint-10000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
38
+ tmp-spec-checkpoint-15000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
39
+ tmp-spec-checkpoint-20000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
40
+ tmp-spec-checkpoint-25000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
41
+ tmp-spec-checkpoint-30000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
42
+ tmp-spec-checkpoint-35000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
43
+ tmp-spec-checkpoint-40000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
44
+ tmp-spec-checkpoint-45000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
45
+ tmp-spec-checkpoint-50000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
46
+ tmp-spec-checkpoint-55000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
47
+ tmp-spec-checkpoint-60000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
48
+ tmp-spec-checkpoint-65000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
49
+ tmp-spec-checkpoint-70000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
50
+ tmp-spec-checkpoint-75000/generated_smiles_1M.csv filter=lfs diff=lfs merge=lfs -text
51
+ tmp-spec-checkpoint-75000/result_unique@1k_unique@10k_IntDiv_IntDiv2_filters_FCD_SNN_Scaf_Frag_logP_SA_QED_weight_NP_NumRings_Bertz_TPSA_AliphaticRings_AromaticRings_RotatableBonds_175k_100k.json filter=lfs diff=lfs merge=lfs -text
52
+ tmp-spec-checkpoint-75000/result_unique@1k_unique@10k_IntDiv_IntDiv2_filters_FCD_SNN_Scaf_Frag_logP_SA_QED_weight_NP_NumRings_Bertz_TPSA_AliphaticRings_AromaticRings_RotatableBonds_500k_100k.json filter=lfs diff=lfs merge=lfs -text
53
+ tmp-spec-checkpoint-75000/result_unique@1k_unique@10k_IntDiv_IntDiv2_filters_FCD_SNN_Scaf_Frag_logP_SA_QED_weight_NP_NumRings_Bertz_TPSA_AliphaticRings_AromaticRings_RotatableBonds_1M_100k.json filter=lfs diff=lfs merge=lfs -text
54
+ tmp-spec-checkpoint-75000/result_unique@1k_unique@10k_IntDiv_IntDiv2_filters_FCD_SNN_Scaf_Frag_logP_SA_QED_weight_NP_NumRings_Bertz_TPSA_AliphaticRings_AromaticRings_RotatableBonds_3M_100k.json filter=lfs diff=lfs merge=lfs -text
55
+ tmp-spec-checkpoint-5000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
56
+ tmp-spec-checkpoint-10000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
57
+ tmp-spec-checkpoint-15000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
58
+ tmp-spec-checkpoint-20000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
59
+ tmp-spec-checkpoint-25000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
60
+ tmp-spec-checkpoint-30000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
61
+ tmp-spec-checkpoint-35000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
62
+ tmp-spec-checkpoint-40000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
63
+ tmp-spec-checkpoint-45000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
64
+ tmp-spec-checkpoint-50000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
65
+ tmp-spec-checkpoint-55000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
66
+ tmp-spec-checkpoint-60000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
67
+ tmp-spec-checkpoint-65000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
68
+ tmp-spec-checkpoint-70000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
69
+ tmp-spec-checkpoint-75000/generated_smiles_1M_0.csv filter=lfs diff=lfs merge=lfs -text
70
+ tmp-spec-checkpoint-5000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
71
+ tmp-spec-checkpoint-10000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
72
+ tmp-spec-checkpoint-15000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
73
+ tmp-spec-checkpoint-20000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
74
+ tmp-spec-checkpoint-25000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
75
+ tmp-spec-checkpoint-30000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
76
+ tmp-spec-checkpoint-35000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
77
+ tmp-spec-checkpoint-40000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
78
+ tmp-spec-checkpoint-45000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
79
+ tmp-spec-checkpoint-50000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
80
+ tmp-spec-checkpoint-55000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
81
+ tmp-spec-checkpoint-60000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
82
+ tmp-spec-checkpoint-65000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
83
+ tmp-spec-checkpoint-70000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
84
+ tmp-spec-checkpoint-75000/generated_smiles_1M_1.csv filter=lfs diff=lfs merge=lfs -text
85
+ tmp-spec-checkpoint-5000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
86
+ tmp-spec-checkpoint-10000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
87
+ tmp-spec-checkpoint-15000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
88
+ tmp-spec-checkpoint-20000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
89
+ tmp-spec-checkpoint-25000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
90
+ tmp-spec-checkpoint-30000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
91
+ tmp-spec-checkpoint-35000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
92
+ tmp-spec-checkpoint-40000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
93
+ tmp-spec-checkpoint-45000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
94
+ tmp-spec-checkpoint-50000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
95
+ tmp-spec-checkpoint-55000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
96
+ tmp-spec-checkpoint-60000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
97
+ tmp-spec-checkpoint-65000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
98
+ tmp-spec-checkpoint-70000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
99
+ tmp-spec-checkpoint-75000/generated_smiles_1M_2.csv filter=lfs diff=lfs merge=lfs -text
100
+ tmp-spec-checkpoint-75000/result_unique@1k_unique@10k_IntDiv_IntDiv2_filters_FCD_SNN_Scaf_Frag_logP_SA_QED_weight_NP_NumRings_Bertz_TPSA_AliphaticRings_AromaticRings_RotatableBonds_175k_100k_0.json filter=lfs diff=lfs merge=lfs -text
101
+ tmp-spec-checkpoint-75000/result_unique@1k_unique@10k_IntDiv_IntDiv2_filters_FCD_SNN_Scaf_Frag_logP_SA_QED_weight_NP_NumRings_Bertz_TPSA_AliphaticRings_AromaticRings_RotatableBonds_1M_100k_0.json filter=lfs diff=lfs merge=lfs -text
102
+ tmp-spec-checkpoint-75000/result_unique@1k_unique@10k_IntDiv_IntDiv2_filters_FCD_SNN_Scaf_Frag_logP_SA_QED_weight_NP_NumRings_Bertz_TPSA_AliphaticRings_AromaticRings_RotatableBonds_175k_100k_1.json filter=lfs diff=lfs merge=lfs -text
103
+ tmp-spec-checkpoint-75000/result_unique@1k_unique@10k_IntDiv_IntDiv2_filters_FCD_SNN_Scaf_Frag_logP_SA_QED_weight_NP_NumRings_Bertz_TPSA_AliphaticRings_AromaticRings_RotatableBonds_175k_100k_2.json filter=lfs diff=lfs merge=lfs -text
104
+ tmp-spec-checkpoint-75000/result_unique@1k_unique@10k_IntDiv_IntDiv2_filters_FCD_SNN_Scaf_Frag_logP_SA_QED_weight_NP_NumRings_Bertz_TPSA_AliphaticRings_AromaticRings_RotatableBonds_1M_100k_1.json filter=lfs diff=lfs merge=lfs -text
105
+ tmp-spec-checkpoint-75000/result_unique@1k_unique@10k_IntDiv_IntDiv2_filters_FCD_SNN_Scaf_Frag_logP_SA_QED_weight_NP_NumRings_Bertz_TPSA_AliphaticRings_AromaticRings_RotatableBonds_1M_100k_2.json filter=lfs diff=lfs merge=lfs -text
run_history.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-10000/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attention_bias": false,
3
+ "attention_dropout": 0.0,
4
+ "bos_token_id": 2,
5
+ "eos_token_id": 3,
6
+ "fused_bias_fc": false,
7
+ "fused_dropout_add_ln": false,
8
+ "fused_mlp": false,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 512,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 1024,
13
+ "max_position_embeddings": 2048,
14
+ "max_seq_length": 128,
15
+ "mlp_bias": false,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 8,
18
+ "num_hidden_layers": 12,
19
+ "num_key_value_heads": 8,
20
+ "pretraining_tp": 1,
21
+ "residual_in_fp32": true,
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_scaling": null,
24
+ "rope_theta": 10000.0,
25
+ "tie_word_embeddings": false,
26
+ "transformers_version": "4.43.4",
27
+ "use_cache": true,
28
+ "use_flash_attn": true,
29
+ "vocab_size": 74
30
+ }
tmp-spec-checkpoint-10000/generated_smiles.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-10000/generated_smiles_100k.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-10000/generated_smiles_100k_0.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-10000/generated_smiles_100k_1.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-10000/generated_smiles_100k_2.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-10000/generated_smiles_1M.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e934f218777b44d9fd97b365133f58d2d3619bf0ce88ecafe19371fa7208c097
3
+ size 47094848
tmp-spec-checkpoint-10000/generated_smiles_1M_0.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:731ed4e88cfa6047580d7ad90edbecbce18285785693b55b5c63e4bc46da3a56
3
+ size 47089287
tmp-spec-checkpoint-10000/generated_smiles_1M_1.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d69d253f5cdb360207fc8540d4b0dec98dcdc1fafcf7da3a1b007eb1bd95557a
3
+ size 47084395
tmp-spec-checkpoint-10000/generated_smiles_1M_2.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdb2bae6e87ae6ee5cf4ea15ff03f67cfb0fd67b129cd9f4865a8e9f121d77bd
3
+ size 47057693
tmp-spec-checkpoint-10000/generated_smiles_30k.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-10000/generated_smiles_30k_0.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-10000/generated_smiles_30k_1.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-10000/generated_smiles_30k_2.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-10000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b23fb00abc522e12ddd7ab4a3a81859e4b4208927bea9b1f002540a9e97ec865
3
+ size 63115855
tmp-spec-checkpoint-10000/result_FCD_SNN_Frag_Scaf.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "validity": 1.0,
3
+ "FCD": {
4
+ "FCD": 1.014371570941769,
5
+ "FCD/test": 1.2840516297559006
6
+ },
7
+ "SNN": {
8
+ "SNN": 0.5403234519492657,
9
+ "SNN/test": 0.5177344154944686
10
+ },
11
+ "Frag": {
12
+ "Frag": 0.9954628107892446,
13
+ "Frag/test": 0.9932640709949498
14
+ },
15
+ "Scaf": {
16
+ "Scaf": 0.6604473255551877,
17
+ "Scaf/test": 0.013251719284644577
18
+ }
19
+ }
tmp-spec-checkpoint-10000/result_SA_wasserstein.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "validity": 1.0,
3
+ "SA_wasserstein": {
4
+ "SA_wasserstein": 0.03531485684710113,
5
+ "SA_wasserstein/test": 0.0633226642649299
6
+ }
7
+ }
tmp-spec-checkpoint-10000/result_novelty_total.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "novelty_total": 0.998974949008943
3
+ }
tmp-spec-checkpoint-10000/result_unique@1k_unique@10k_IntDiv_IntDiv2_logP_wasserstein_QED_wasserstein_weight_wasserstein_NP_wasserstein.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "validity": 1.0,
3
+ "unique@1k": 1.0,
4
+ "unique@10k": 1.0,
5
+ "IntDiv": 0.853450210055646,
6
+ "IntDiv2": 0.8493021878803156,
7
+ "logP_wasserstein": {
8
+ "logP_wasserstein": 0.13093312964571951,
9
+ "logP_wasserstein/test": 0.10780172952943484
10
+ },
11
+ "QED_wasserstein": {
12
+ "QED_wasserstein": 0.005999451053057837,
13
+ "QED_wasserstein/test": 0.008760303916285503
14
+ },
15
+ "weight_wasserstein": {
16
+ "weight_wasserstein": 3.6379066160855205,
17
+ "weight_wasserstein/test": 5.916868278108168
18
+ },
19
+ "NP_wasserstein": {
20
+ "NP_wasserstein": 0.049366192344106656,
21
+ "NP_wasserstein/test": 0.07572631593283732
22
+ }
23
+ }
tmp-spec-checkpoint-10000/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<bos>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<eos>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tmp-spec-checkpoint-10000/tokenizer.json ADDED
@@ -0,0 +1,193 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [
6
+ {
7
+ "id": 0,
8
+ "content": "<unk>",
9
+ "single_word": false,
10
+ "lstrip": false,
11
+ "rstrip": false,
12
+ "normalized": false,
13
+ "special": true
14
+ },
15
+ {
16
+ "id": 1,
17
+ "content": "<pad>",
18
+ "single_word": false,
19
+ "lstrip": false,
20
+ "rstrip": false,
21
+ "normalized": false,
22
+ "special": true
23
+ },
24
+ {
25
+ "id": 2,
26
+ "content": "<bos>",
27
+ "single_word": false,
28
+ "lstrip": false,
29
+ "rstrip": false,
30
+ "normalized": false,
31
+ "special": true
32
+ },
33
+ {
34
+ "id": 3,
35
+ "content": "<eos>",
36
+ "single_word": false,
37
+ "lstrip": false,
38
+ "rstrip": false,
39
+ "normalized": false,
40
+ "special": true
41
+ }
42
+ ],
43
+ "normalizer": null,
44
+ "pre_tokenizer": {
45
+ "type": "Split",
46
+ "pattern": {
47
+ "Regex": "(\\[[^\\]]+]|Br?|Cl?|N|O|S|P|F|I|b|c|n|o|s|p|\\(|\\)|\\.|=|#|-|\\+|\\\\\\\\|\\/|:|~|@|\\?|>>?|\\*|\\$|\\%[0-9]{2}|[0-9])"
48
+ },
49
+ "behavior": "Isolated",
50
+ "invert": false
51
+ },
52
+ "post_processor": {
53
+ "type": "TemplateProcessing",
54
+ "single": [
55
+ {
56
+ "SpecialToken": {
57
+ "id": "<bos>",
58
+ "type_id": 0
59
+ }
60
+ },
61
+ {
62
+ "Sequence": {
63
+ "id": "A",
64
+ "type_id": 0
65
+ }
66
+ },
67
+ {
68
+ "SpecialToken": {
69
+ "id": "<eos>",
70
+ "type_id": 0
71
+ }
72
+ }
73
+ ],
74
+ "pair": [
75
+ {
76
+ "Sequence": {
77
+ "id": "A",
78
+ "type_id": 0
79
+ }
80
+ },
81
+ {
82
+ "Sequence": {
83
+ "id": "B",
84
+ "type_id": 1
85
+ }
86
+ }
87
+ ],
88
+ "special_tokens": {
89
+ "<bos>": {
90
+ "id": "<bos>",
91
+ "ids": [
92
+ 2
93
+ ],
94
+ "tokens": [
95
+ "<bos>"
96
+ ]
97
+ },
98
+ "<eos>": {
99
+ "id": "<eos>",
100
+ "ids": [
101
+ 3
102
+ ],
103
+ "tokens": [
104
+ "<eos>"
105
+ ]
106
+ }
107
+ }
108
+ },
109
+ "decoder": {
110
+ "type": "BPEDecoder",
111
+ "suffix": "</w>"
112
+ },
113
+ "model": {
114
+ "type": "WordLevel",
115
+ "vocab": {
116
+ "<unk>": 0,
117
+ "<pad>": 1,
118
+ "<bos>": 2,
119
+ "<eos>": 3,
120
+ "C": 4,
121
+ ".": 5,
122
+ "1": 6,
123
+ "c": 7,
124
+ "O": 8,
125
+ "=": 9,
126
+ "(": 10,
127
+ ")": 11,
128
+ "N": 12,
129
+ "5": 13,
130
+ "6": 14,
131
+ "4": 15,
132
+ "7": 16,
133
+ "8": 17,
134
+ "3": 18,
135
+ "2": 19,
136
+ "n": 20,
137
+ "9": 21,
138
+ "F": 22,
139
+ "%10": 23,
140
+ "%11": 24,
141
+ "S": 25,
142
+ "s": 26,
143
+ "Cl": 27,
144
+ "[nH]": 28,
145
+ "o": 29,
146
+ "#": 30,
147
+ "Br": 31,
148
+ "%12": 32,
149
+ "[N+]": 33,
150
+ "[O-]": 34,
151
+ "%13": 35,
152
+ "I": 36,
153
+ "[N-]": 37,
154
+ "P": 38,
155
+ "[n+]": 39,
156
+ "%14": 40,
157
+ "-": 41,
158
+ "[Si]": 42,
159
+ "[S+]": 43,
160
+ "%15": 44,
161
+ "B": 45,
162
+ "%16": 46,
163
+ "[NH+]": 47,
164
+ "[B-]": 48,
165
+ "%17": 49,
166
+ "[NH2+]": 50,
167
+ "[O]": 51,
168
+ "[NH3+]": 52,
169
+ "[PH]": 53,
170
+ "[n-]": 54,
171
+ "%18": 55,
172
+ "[nH+]": 56,
173
+ "[Sn]": 57,
174
+ "[s+]": 58,
175
+ "%19": 59,
176
+ "[Se]": 60,
177
+ "[Cl-]": 61,
178
+ "%20": 62,
179
+ "[N]": 63,
180
+ "[C-]": 64,
181
+ "[C]": 65,
182
+ "[SiH]": 66,
183
+ "%21": 67,
184
+ "[O+]": 68,
185
+ "[SH]": 69,
186
+ "[NH]": 70,
187
+ "[P+]": 71,
188
+ "[c-]": 72,
189
+ "[o+]": 73
190
+ },
191
+ "unk_token": "<unk>"
192
+ }
193
+ }
tmp-spec-checkpoint-10000/tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "<bos>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<eos>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ }
35
+ },
36
+ "bos_token": "<bos>",
37
+ "clean_up_tokenization_spaces": true,
38
+ "eos_token": "<eos>",
39
+ "model_max_length": 1000000000000000019884624838656,
40
+ "pad_token": "<pad>",
41
+ "tokenizer_class": "PreTrainedTokenizerFast",
42
+ "unk_token": "<unk>"
43
+ }
tmp-spec-checkpoint-10000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:634148355b42ee0de306ebb5fc5e0474a7658df467b360eba0c60e507cf38172
3
+ size 6584
tmp-spec-checkpoint-15000/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attention_bias": false,
3
+ "attention_dropout": 0.0,
4
+ "bos_token_id": 2,
5
+ "eos_token_id": 3,
6
+ "fused_bias_fc": false,
7
+ "fused_dropout_add_ln": false,
8
+ "fused_mlp": false,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 512,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 1024,
13
+ "max_position_embeddings": 2048,
14
+ "max_seq_length": 128,
15
+ "mlp_bias": false,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 8,
18
+ "num_hidden_layers": 12,
19
+ "num_key_value_heads": 8,
20
+ "pretraining_tp": 1,
21
+ "residual_in_fp32": true,
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_scaling": null,
24
+ "rope_theta": 10000.0,
25
+ "tie_word_embeddings": false,
26
+ "transformers_version": "4.43.4",
27
+ "use_cache": true,
28
+ "use_flash_attn": true,
29
+ "vocab_size": 74
30
+ }
tmp-spec-checkpoint-15000/generated_smiles.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-15000/generated_smiles_100k.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-15000/generated_smiles_100k_0.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-15000/generated_smiles_100k_1.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-15000/generated_smiles_100k_2.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-15000/generated_smiles_1M.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2232d0bab162da91fc2180f6d75f8aeb8619ffca8e72bdac100b3ffa65d83295
3
+ size 47298381
tmp-spec-checkpoint-15000/generated_smiles_1M_0.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe29ffbc52aa4bf86ea6a418fb1a04927771efccf191d5fa6500d38161f1707d
3
+ size 47315469
tmp-spec-checkpoint-15000/generated_smiles_1M_1.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f52211d12daca2fe0e55661f719a078aab322850a36cf387b38ce69ac4621908
3
+ size 47330945
tmp-spec-checkpoint-15000/generated_smiles_1M_2.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3af8c0201ed464d343c7dab853b3e7e525d0111e88a7e2cf21a6968e64681851
3
+ size 47298298
tmp-spec-checkpoint-15000/generated_smiles_30k.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-15000/generated_smiles_30k_0.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-15000/generated_smiles_30k_1.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-15000/generated_smiles_30k_2.csv ADDED
The diff for this file is too large to render. See raw diff
 
tmp-spec-checkpoint-15000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b1d86a6466f9eaa4f20e70229d2fed4e81e52000a8a334a64b55b1b05ef70a4
3
+ size 63115855
tmp-spec-checkpoint-15000/result_FCD_SNN_Frag_Scaf.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "validity": 1.0,
3
+ "FCD": {
4
+ "FCD": 1.002989927399483,
5
+ "FCD/test": 1.2865087128018615
6
+ },
7
+ "SNN": {
8
+ "SNN": 0.5417524048349963,
9
+ "SNN/test": 0.5186626958462258
10
+ },
11
+ "Frag": {
12
+ "Frag": 0.9954620308507153,
13
+ "Frag/test": 0.9930898868011179
14
+ },
15
+ "Scaf": {
16
+ "Scaf": 0.6705683290696588,
17
+ "Scaf/test": 0.008874388198630616
18
+ }
19
+ }
tmp-spec-checkpoint-15000/result_SA_wasserstein.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "validity": 1.0,
3
+ "SA_wasserstein": {
4
+ "SA_wasserstein": 0.026378598518104434,
5
+ "SA_wasserstein/test": 0.054440884921422565
6
+ }
7
+ }
tmp-spec-checkpoint-15000/result_novelty_total.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "novelty_total": 0.9991406413749738
3
+ }
tmp-spec-checkpoint-15000/result_unique@1k_unique@10k_IntDiv_IntDiv2_logP_wasserstein_QED_wasserstein_weight_wasserstein_NP_wasserstein.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "validity": 1.0,
3
+ "unique@1k": 1.0,
4
+ "unique@10k": 1.0,
5
+ "IntDiv": 0.8539086081502472,
6
+ "IntDiv2": 0.8497642859746146,
7
+ "logP_wasserstein": {
8
+ "logP_wasserstein": 0.1386001593623254,
9
+ "logP_wasserstein/test": 0.11556755788981397
10
+ },
11
+ "QED_wasserstein": {
12
+ "QED_wasserstein": 0.005218756412930638,
13
+ "QED_wasserstein/test": 0.009414825617739903
14
+ },
15
+ "weight_wasserstein": {
16
+ "weight_wasserstein": 2.698026518070816,
17
+ "weight_wasserstein/test": 4.977317290493286
18
+ },
19
+ "NP_wasserstein": {
20
+ "NP_wasserstein": 0.046379309027314096,
21
+ "NP_wasserstein/test": 0.07274505519831714
22
+ }
23
+ }
tmp-spec-checkpoint-15000/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<bos>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<eos>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tmp-spec-checkpoint-15000/tokenizer.json ADDED
@@ -0,0 +1,193 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [
6
+ {
7
+ "id": 0,
8
+ "content": "<unk>",
9
+ "single_word": false,
10
+ "lstrip": false,
11
+ "rstrip": false,
12
+ "normalized": false,
13
+ "special": true
14
+ },
15
+ {
16
+ "id": 1,
17
+ "content": "<pad>",
18
+ "single_word": false,
19
+ "lstrip": false,
20
+ "rstrip": false,
21
+ "normalized": false,
22
+ "special": true
23
+ },
24
+ {
25
+ "id": 2,
26
+ "content": "<bos>",
27
+ "single_word": false,
28
+ "lstrip": false,
29
+ "rstrip": false,
30
+ "normalized": false,
31
+ "special": true
32
+ },
33
+ {
34
+ "id": 3,
35
+ "content": "<eos>",
36
+ "single_word": false,
37
+ "lstrip": false,
38
+ "rstrip": false,
39
+ "normalized": false,
40
+ "special": true
41
+ }
42
+ ],
43
+ "normalizer": null,
44
+ "pre_tokenizer": {
45
+ "type": "Split",
46
+ "pattern": {
47
+ "Regex": "(\\[[^\\]]+]|Br?|Cl?|N|O|S|P|F|I|b|c|n|o|s|p|\\(|\\)|\\.|=|#|-|\\+|\\\\\\\\|\\/|:|~|@|\\?|>>?|\\*|\\$|\\%[0-9]{2}|[0-9])"
48
+ },
49
+ "behavior": "Isolated",
50
+ "invert": false
51
+ },
52
+ "post_processor": {
53
+ "type": "TemplateProcessing",
54
+ "single": [
55
+ {
56
+ "SpecialToken": {
57
+ "id": "<bos>",
58
+ "type_id": 0
59
+ }
60
+ },
61
+ {
62
+ "Sequence": {
63
+ "id": "A",
64
+ "type_id": 0
65
+ }
66
+ },
67
+ {
68
+ "SpecialToken": {
69
+ "id": "<eos>",
70
+ "type_id": 0
71
+ }
72
+ }
73
+ ],
74
+ "pair": [
75
+ {
76
+ "Sequence": {
77
+ "id": "A",
78
+ "type_id": 0
79
+ }
80
+ },
81
+ {
82
+ "Sequence": {
83
+ "id": "B",
84
+ "type_id": 1
85
+ }
86
+ }
87
+ ],
88
+ "special_tokens": {
89
+ "<bos>": {
90
+ "id": "<bos>",
91
+ "ids": [
92
+ 2
93
+ ],
94
+ "tokens": [
95
+ "<bos>"
96
+ ]
97
+ },
98
+ "<eos>": {
99
+ "id": "<eos>",
100
+ "ids": [
101
+ 3
102
+ ],
103
+ "tokens": [
104
+ "<eos>"
105
+ ]
106
+ }
107
+ }
108
+ },
109
+ "decoder": {
110
+ "type": "BPEDecoder",
111
+ "suffix": "</w>"
112
+ },
113
+ "model": {
114
+ "type": "WordLevel",
115
+ "vocab": {
116
+ "<unk>": 0,
117
+ "<pad>": 1,
118
+ "<bos>": 2,
119
+ "<eos>": 3,
120
+ "C": 4,
121
+ ".": 5,
122
+ "1": 6,
123
+ "c": 7,
124
+ "O": 8,
125
+ "=": 9,
126
+ "(": 10,
127
+ ")": 11,
128
+ "N": 12,
129
+ "5": 13,
130
+ "6": 14,
131
+ "4": 15,
132
+ "7": 16,
133
+ "8": 17,
134
+ "3": 18,
135
+ "2": 19,
136
+ "n": 20,
137
+ "9": 21,
138
+ "F": 22,
139
+ "%10": 23,
140
+ "%11": 24,
141
+ "S": 25,
142
+ "s": 26,
143
+ "Cl": 27,
144
+ "[nH]": 28,
145
+ "o": 29,
146
+ "#": 30,
147
+ "Br": 31,
148
+ "%12": 32,
149
+ "[N+]": 33,
150
+ "[O-]": 34,
151
+ "%13": 35,
152
+ "I": 36,
153
+ "[N-]": 37,
154
+ "P": 38,
155
+ "[n+]": 39,
156
+ "%14": 40,
157
+ "-": 41,
158
+ "[Si]": 42,
159
+ "[S+]": 43,
160
+ "%15": 44,
161
+ "B": 45,
162
+ "%16": 46,
163
+ "[NH+]": 47,
164
+ "[B-]": 48,
165
+ "%17": 49,
166
+ "[NH2+]": 50,
167
+ "[O]": 51,
168
+ "[NH3+]": 52,
169
+ "[PH]": 53,
170
+ "[n-]": 54,
171
+ "%18": 55,
172
+ "[nH+]": 56,
173
+ "[Sn]": 57,
174
+ "[s+]": 58,
175
+ "%19": 59,
176
+ "[Se]": 60,
177
+ "[Cl-]": 61,
178
+ "%20": 62,
179
+ "[N]": 63,
180
+ "[C-]": 64,
181
+ "[C]": 65,
182
+ "[SiH]": 66,
183
+ "%21": 67,
184
+ "[O+]": 68,
185
+ "[SH]": 69,
186
+ "[NH]": 70,
187
+ "[P+]": 71,
188
+ "[c-]": 72,
189
+ "[o+]": 73
190
+ },
191
+ "unk_token": "<unk>"
192
+ }
193
+ }
tmp-spec-checkpoint-15000/tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "<bos>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<eos>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ }
35
+ },
36
+ "bos_token": "<bos>",
37
+ "clean_up_tokenization_spaces": true,
38
+ "eos_token": "<eos>",
39
+ "model_max_length": 1000000000000000019884624838656,
40
+ "pad_token": "<pad>",
41
+ "tokenizer_class": "PreTrainedTokenizerFast",
42
+ "unk_token": "<unk>"
43
+ }
tmp-spec-checkpoint-15000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:634148355b42ee0de306ebb5fc5e0474a7658df467b360eba0c60e507cf38172
3
+ size 6584
tmp-spec-checkpoint-20000/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attention_bias": false,
3
+ "attention_dropout": 0.0,
4
+ "bos_token_id": 2,
5
+ "eos_token_id": 3,
6
+ "fused_bias_fc": false,
7
+ "fused_dropout_add_ln": false,
8
+ "fused_mlp": false,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 512,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 1024,
13
+ "max_position_embeddings": 2048,
14
+ "max_seq_length": 128,
15
+ "mlp_bias": false,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 8,
18
+ "num_hidden_layers": 12,
19
+ "num_key_value_heads": 8,
20
+ "pretraining_tp": 1,
21
+ "residual_in_fp32": true,
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_scaling": null,
24
+ "rope_theta": 10000.0,
25
+ "tie_word_embeddings": false,
26
+ "transformers_version": "4.43.4",
27
+ "use_cache": true,
28
+ "use_flash_attn": true,
29
+ "vocab_size": 74
30
+ }
tmp-spec-checkpoint-20000/generated_smiles.csv ADDED
The diff for this file is too large to render. See raw diff