Update README.md
Browse files
README.md
CHANGED
|
@@ -43,15 +43,15 @@ inference:
|
|
| 43 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 44 |
should probably proofread and complete it, then remove this comment. -->
|
| 45 |
|
| 46 |
-
# distilgpt2-
|
| 47 |
|
| 48 |
-
This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on
|
| 49 |
It achieves the following results on the evaluation set:
|
| 50 |
- Loss: 2.2461
|
| 51 |
|
| 52 |
## Model description
|
| 53 |
|
| 54 |
-
|
| 55 |
|
| 56 |
## Intended uses & limitations
|
| 57 |
|
|
@@ -63,6 +63,7 @@ More information needed
|
|
| 63 |
|
| 64 |
## Training procedure
|
| 65 |
|
|
|
|
| 66 |
### Training hyperparameters
|
| 67 |
|
| 68 |
The following hyperparameters were used during training:
|
|
|
|
| 43 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 44 |
should probably proofread and complete it, then remove this comment. -->
|
| 45 |
|
| 46 |
+
# distilgpt2-tiny-conversational
|
| 47 |
|
| 48 |
+
This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on a parsed version of Wizard of Wikipedia. Persona alpha/beta framework designed for use with [ai-msgbot](https://github.com/pszemraj/ai-msgbot).
|
| 49 |
It achieves the following results on the evaluation set:
|
| 50 |
- Loss: 2.2461
|
| 51 |
|
| 52 |
## Model description
|
| 53 |
|
| 54 |
+
- it is a splendid model
|
| 55 |
|
| 56 |
## Intended uses & limitations
|
| 57 |
|
|
|
|
| 63 |
|
| 64 |
## Training procedure
|
| 65 |
|
| 66 |
+
- deepspeed
|
| 67 |
### Training hyperparameters
|
| 68 |
|
| 69 |
The following hyperparameters were used during training:
|