ribesstefano commited on
Commit
25e2f0f
·
verified ·
1 Parent(s): 167e1f2

Initial version

Browse files
README.md ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: mit
4
+ base_model: seyonec/ChemBERTa-zinc-base-v1
5
+ tags:
6
+ - PROTAC
7
+ - cheminformatics
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: ailab-bio/PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25-rand-smiles
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # ailab-bio/PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25-rand-smiles
18
+
19
+ This model is a fine-tuned version of [seyonec/ChemBERTa-zinc-base-v1](https://huggingface.co/seyonec/ChemBERTa-zinc-base-v1) on the ailab-bio/PROTAC-Splitter-Dataset dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.3184
22
+ - E3 Tanimoto Similarity: 0.0
23
+ - Poi Equal: 0.7620
24
+ - E3 Equal: 0.8036
25
+ - Poi Tanimoto Similarity: 0.0
26
+ - Poi Valid: 0.9589
27
+ - Reassembly: 0.5459
28
+ - All Ligands Equal: 0.5390
29
+ - Poi Graph Edit Distance: inf
30
+ - Linker Graph Edit Distance: 28328611898016997512352231618590754490884636453008387363307520.0000
31
+ - Poi Graph Edit Distance Norm: inf
32
+ - Linker Tanimoto Similarity: 0.0
33
+ - Valid: 0.9547
34
+ - E3 Graph Edit Distance Norm: inf
35
+ - Poi Has Attachment Point(s): 0.9589
36
+ - Linker Equal: 0.7668
37
+ - Linker Heavy Atoms Difference Norm: 0.0034
38
+ - Has All Attachment Points: 0.9905
39
+ - Heavy Atoms Difference: 4.1049
40
+ - Has Three Substructures: 0.9992
41
+ - Tanimoto Similarity: 0.0
42
+ - Heavy Atoms Difference Norm: 0.0536
43
+ - Poi Heavy Atoms Difference Norm: 0.0352
44
+ - Linker Has Attachment Point(s): 0.9972
45
+ - E3 Heavy Atoms Difference: 0.2976
46
+ - E3 Graph Edit Distance: inf
47
+ - Linker Valid: 0.9972
48
+ - E3 Heavy Atoms Difference Norm: 0.0033
49
+ - E3 Has Attachment Point(s): 0.9966
50
+ - E3 Valid: 0.9966
51
+ - Reassembly Nostereo: 0.5799
52
+ - Linker Heavy Atoms Difference: 0.3031
53
+ - Num Fragments: 3.0003
54
+ - Linker Graph Edit Distance Norm: inf
55
+ - Poi Heavy Atoms Difference: 1.1856
56
+
57
+ ## Model description
58
+
59
+ More information needed
60
+
61
+ ## Intended uses & limitations
62
+
63
+ More information needed
64
+
65
+ ## Training and evaluation data
66
+
67
+ More information needed
68
+
69
+ ## Training procedure
70
+
71
+ ### Training hyperparameters
72
+
73
+ The following hyperparameters were used during training:
74
+ - learning_rate: 5e-05
75
+ - train_batch_size: 128
76
+ - eval_batch_size: 64
77
+ - seed: 42
78
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
79
+ - lr_scheduler_type: cosine
80
+ - lr_scheduler_warmup_steps: 699
81
+ - training_steps: 10000
82
+ - mixed_precision_training: Native AMP
83
+
84
+ ### Training results
85
+
86
+ | Training Loss | Epoch | Step | Validation Loss | E3 Tanimoto Similarity | Poi Equal | E3 Equal | Poi Tanimoto Similarity | Poi Valid | Reassembly | All Ligands Equal | Poi Graph Edit Distance | Linker Graph Edit Distance | Poi Graph Edit Distance Norm | Linker Tanimoto Similarity | Valid | E3 Graph Edit Distance Norm | Poi Has Attachment Point(s) | Linker Equal | Linker Heavy Atoms Difference Norm | Has All Attachment Points | Heavy Atoms Difference | Has Three Substructures | Tanimoto Similarity | Heavy Atoms Difference Norm | Poi Heavy Atoms Difference Norm | Linker Has Attachment Point(s) | E3 Heavy Atoms Difference | E3 Graph Edit Distance | Linker Valid | E3 Heavy Atoms Difference Norm | E3 Has Attachment Point(s) | E3 Valid | Reassembly Nostereo | Linker Heavy Atoms Difference | Num Fragments | Linker Graph Edit Distance Norm | Poi Heavy Atoms Difference |
87
+ |:-------------:|:------:|:-----:|:---------------:|:----------------------:|:---------:|:--------:|:-----------------------:|:---------:|:----------:|:-----------------:|:-----------------------:|:-------------------------------------------------------------------:|:----------------------------:|:--------------------------:|:------:|:---------------------------:|:---------------------------:|:------------:|:----------------------------------:|:-------------------------:|:----------------------:|:-----------------------:|:-------------------:|:---------------------------:|:-------------------------------:|:------------------------------:|:-------------------------:|:----------------------:|:------------:|:------------------------------:|:--------------------------:|:--------:|:-------------------:|:-----------------------------:|:-------------:|:-------------------------------:|:--------------------------:|
88
+ | 0.012 | 0.4932 | 5000 | 0.3055 | 0.0 | 0.7354 | 0.7897 | 0.0 | 0.9479 | 0.4899 | 0.4824 | inf | 25672804532577903995569209904347871257364201785538851047997440.0000 | inf | 0.0 | 0.9441 | inf | 0.9479 | 0.6946 | 0.0103 | 0.9873 | 4.9586 | 0.9995 | 0.0 | 0.0650 | 0.0421 | 0.9974 | 0.1648 | inf | 0.9974 | -0.0036 | 0.9972 | 0.9972 | 0.5198 | 0.5344 | 3.0005 | inf | 1.4326 |
89
+ | 0.0064 | 0.7398 | 7500 | 0.3156 | 0.0 | 0.7551 | 0.8017 | 0.0 | 0.9532 | 0.5359 | 0.5282 | inf | 40722379603399433924006332951724209580646664901199556834754560.0000 | inf | 0.0 | 0.9477 | inf | 0.9532 | 0.7533 | -0.0000 | 0.9873 | 4.5195 | 0.9985 | 0.0 | 0.0593 | 0.0396 | 0.9959 | 0.3429 | inf | 0.9959 | 0.0053 | 0.9944 | 0.9944 | 0.5689 | 0.2564 | 3.0003 | 0.0452 | 1.3395 |
90
+ | 0.0051 | 0.9864 | 10000 | 0.3184 | 0.0 | 0.7620 | 0.8036 | 0.0 | 0.9589 | 0.5459 | 0.5390 | inf | 28328611898016997512352231618590754490884636453008387363307520.0000 | inf | 0.0 | 0.9547 | inf | 0.9589 | 0.7668 | 0.0034 | 0.9905 | 4.1049 | 0.9992 | 0.0 | 0.0536 | 0.0352 | 0.9972 | 0.2976 | inf | 0.9972 | 0.0033 | 0.9966 | 0.9966 | 0.5799 | 0.3031 | 3.0003 | inf | 1.1856 |
91
+
92
+
93
+ ### Framework versions
94
+
95
+ - Transformers 4.44.2
96
+ - Pytorch 2.4.1+cu121
97
+ - Datasets 3.0.0
98
+ - Tokenizers 0.19.1
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 0,
3
+ "do_sample": true,
4
+ "eos_token_id": 2,
5
+ "max_length": 512,
6
+ "num_beams": 5,
7
+ "pad_token_id": 1,
8
+ "top_k": 20,
9
+ "transformers_version": "4.44.2"
10
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3351ddf7c4d2470694590f4fcc5dca649e784a9ecf2f67f0bf3326d7bbafeaf6
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6b9fcea6d6d61860f983703a2d65e182d32b7cf4bd5a62beb9ec8eb364ffe8e
3
  size 409608164
runs/Mar06_21-17-17_alvis3-24/events.out.tfevents.1741292241.alvis3-24.52083.25 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d46a3d5e5daf61a445ee8a2fed5a9f8ff30f0f4924f981ffa247056b65786bf
3
+ size 23409
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bcc44b1f70fa7b60284d01f87f8bae908d705f358c468dce0c892d39cd1ffa9a
3
  size 7480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e516a35e56487ee3000e6ce07e244c74826c1e9427934404d1d2cc6d84aa7dc
3
  size 7480