|
|
--- |
|
|
license: apache-2.0 |
|
|
datasets: |
|
|
- abacusai/MetaMathFewshot |
|
|
--- |
|
|
 |
|
|
|
|
|
DPO finetune of our [MetaMath SFT Model](https://huggingface.co/abacusai/MM-Orc-Vic-bagel-34b-c1000) on the [Truthy DPO dataset](https://huggingface.co/datasets/jondurbin/truthy-dpo-v0.1) |
|
|
|
|
|
### Evaluation Results |
|
|
|
|
|
| Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K | |
|
|
| --- | --- | --- | --- | --- | --- | --- | |
|
|
| 75.54 | 69.20 | 84.34 | 76.46 | 67.58 | 82.87 | 72.78 | |