Update README.md
#1
by
mkshing
- opened
README.md
CHANGED
|
@@ -56,7 +56,7 @@ print(tokenizer.decode(tokens[0], skip_special_tokens=True))
|
|
| 56 |
|
| 57 |
### Training Procedure
|
| 58 |
|
| 59 |
-
Models are pre-trained on the aforementioned dataset in mixed-precision (FP16), optimized with Adam, and trained using the NeoX tokenizer with a vocabulary size of 50,257. We outline the complete hyperparameters choices in the project's [GitHub repository](https://github.com/Stability-AI/StableLM
|
| 60 |
|
| 61 |
## Use and Limitations
|
| 62 |
|
|
|
|
| 56 |
|
| 57 |
### Training Procedure
|
| 58 |
|
| 59 |
+
Models are pre-trained on the aforementioned dataset in mixed-precision (FP16), optimized with Adam, and trained using the NeoX tokenizer with a vocabulary size of 50,257. We outline the complete hyperparameters choices in the project's [GitHub repository](https://github.com/Stability-AI/StableLM/blob/main/configs/stablelm-base-alpha-7b.yaml).
|
| 60 |
|
| 61 |
## Use and Limitations
|
| 62 |
|