Update README.md
Browse files
README.md
CHANGED
|
@@ -8,18 +8,11 @@ tags:
|
|
| 8 |
|
| 9 |
# GPT2 Zinc 87m
|
| 10 |
|
| 11 |
-
This is a GPT2 style autoregressive language model trained on ~480m SMILES strings from the [ZINC database](https://zinc.docking.org/)
|
| 12 |
|
| 13 |
The model has ~87m parameters and was trained for 175000 iterations with a batch size of 3072 to a validation loss of ~.615. This model is useful for generating druglike molecules or generating embeddings from SMILES strings
|
| 14 |
|
| 15 |
## How to use
|
| 16 |
-
To use, install the [transformers](https://github.com/huggingface/transformers) library:
|
| 17 |
-
|
| 18 |
-
```
|
| 19 |
-
pip install transformers
|
| 20 |
-
```
|
| 21 |
-
|
| 22 |
-
Load the model from the Huggingface Hub:
|
| 23 |
|
| 24 |
```python
|
| 25 |
from transformers import GPT2TokenizerFast, GPT2LMHeadModel
|
|
|
|
| 8 |
|
| 9 |
# GPT2 Zinc 87m
|
| 10 |
|
| 11 |
+
This is a GPT2 style autoregressive language model trained on ~480m SMILES strings from the [ZINC database](https://zinc.docking.org/).
|
| 12 |
|
| 13 |
The model has ~87m parameters and was trained for 175000 iterations with a batch size of 3072 to a validation loss of ~.615. This model is useful for generating druglike molecules or generating embeddings from SMILES strings
|
| 14 |
|
| 15 |
## How to use
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 16 |
|
| 17 |
```python
|
| 18 |
from transformers import GPT2TokenizerFast, GPT2LMHeadModel
|