amd olmo 1b set
#3
by
Thushanthiga
- opened
README.md
CHANGED
|
@@ -3,7 +3,6 @@ license: apache-2.0
|
|
| 3 |
datasets:
|
| 4 |
- allenai/dolma
|
| 5 |
pipeline_tag: text-generation
|
| 6 |
-
library_name: transformers
|
| 7 |
---
|
| 8 |
# AMD-OLMo
|
| 9 |
|
|
@@ -283,11 +282,12 @@ hf-align/scripts/run_dpo.py hf-align/recipes/AMD-OLMo-1B-dpo.yaml \
|
|
| 283 |
|
| 284 |
Feel free to cite our AMD-OLMo models:
|
| 285 |
```bash
|
| 286 |
-
@
|
| 287 |
-
|
| 288 |
-
|
| 289 |
-
|
| 290 |
-
|
|
|
|
| 291 |
}
|
| 292 |
```
|
| 293 |
|
|
|
|
| 3 |
datasets:
|
| 4 |
- allenai/dolma
|
| 5 |
pipeline_tag: text-generation
|
|
|
|
| 6 |
---
|
| 7 |
# AMD-OLMo
|
| 8 |
|
|
|
|
| 282 |
|
| 283 |
Feel free to cite our AMD-OLMo models:
|
| 284 |
```bash
|
| 285 |
+
@misc{AMD-OLMo,
|
| 286 |
+
title = {AMD-OLMo: A series of 1B language models trained from scratch by AMD on AMD Instinct™ MI250 GPUs.},
|
| 287 |
+
url = {https://huggingface.co/amd/AMD-OLMo},
|
| 288 |
+
author = {Jiang Liu, Jialian Wu, Prakamya Mishra, Zicheng Liu, Sudhanshu Ranjan, Pratik Prabhanjan Brahma, Yusheng Su, Gowtham Ramesh, Peng Sun, Zhe Li, Dong Li, Lu Tian, Emad Barsoum},
|
| 289 |
+
month = {October},
|
| 290 |
+
year = {2024}
|
| 291 |
}
|
| 292 |
```
|
| 293 |
|