Update README.md
#2
by jburtoft - opened
README.md
CHANGED
|
@@ -9,7 +9,7 @@ This model experiment was inspired by the work published in [Goat: Fine-tuned LL
|
|
| 9 |
|
| 10 |
Fine tuning of [philschmid/Llama-2-7b-hf](https://huggingface.co/philschmid/Llama-2-7b-hf) was conducted with 2.8M math problems from the [AtlasUnified/atlas-math-sets](https://huggingface.co/datasets/AtlasUnified/atlas-math-sets) dataset.
|
| 11 |
|
| 12 |
-
Training was conducted on a trn1.32xlarge instance. The model here was complied for 2 Neuron cores, which will run on AWS inf2.8xlarge and larger instances.
|
| 13 |
|
| 14 |
The table below highlights performance improvements observed on basic math questions.
|
| 15 |
|
|
|
|
| 9 |
|
| 10 |
Fine tuning of [philschmid/Llama-2-7b-hf](https://huggingface.co/philschmid/Llama-2-7b-hf) was conducted with 2.8M math problems from the [AtlasUnified/atlas-math-sets](https://huggingface.co/datasets/AtlasUnified/atlas-math-sets) dataset.
|
| 11 |
|
| 12 |
+
Training was conducted on a trn1.32xlarge instance. The model here was complied for 2 Neuron cores, which will run on AWS inf2.8xlarge and larger instances. It also runs on an inf2.xlarge (the smallest Inferentia2 instance), but it uses over 80% of the system RAM.
|
| 13 |
|
| 14 |
The table below highlights performance improvements observed on basic math questions.
|
| 15 |
|