Update README.md
Browse files
README.md
CHANGED
|
@@ -87,7 +87,7 @@ model = PeftModel.from_pretrained(base_model, "suwesh/llamatron-1B-peft", subfol
|
|
| 87 |
<pre>Checkpoint 11000 Training and Validation losses: 1.06 | 1.09</pre>
|
| 88 |
|
| 89 |
# Evaluation details
|
| 90 |
-
We use the [nvidia/Llama-3.1-Nemotron-Nano](https://huggingface.co/nvidia/Llama-3.1-Nemotron-Nano-8B-v1) LLM as a Judge for evaluating the responses between the base llama 3.2 1b instruct and
|
| 91 |
<pre>base: 122
|
| 92 |
peft: 388
|
| 93 |
tie: 29
|
|
|
|
| 87 |
<pre>Checkpoint 11000 Training and Validation losses: 1.06 | 1.09</pre>
|
| 88 |
|
| 89 |
# Evaluation details
|
| 90 |
+
We use the [nvidia/Llama-3.1-Nemotron-Nano](https://huggingface.co/nvidia/Llama-3.1-Nemotron-Nano-8B-v1) LLM as a Judge for evaluating the responses between the base llama 3.2 1b instruct and our PEFT model. The following are the judge's preference for each prompt to the two models, we also provide the ground truth in the prompt to the judge:
|
| 91 |
<pre>base: 122
|
| 92 |
peft: 388
|
| 93 |
tie: 29
|