| | --- |
| | license: apache-2.0 |
| | datasets: |
| | - EleutherAI/muInstruct |
| | - camel-ai/math |
| | language: |
| | - en |
| | tags: |
| | - math |
| | --- |
| | |
| | `llemma_7b_muinstruct_camelmath` is an instruction-following finetune of [Llemma 7B](https://huggingface.co/EleutherAI/llemma_7b), trained on the [μInstruct](https://huggingface.co/datasets/EleutherAI/muInstruct) and [camel-ai/math](https://huggingface.co/datasets/camel-ai/math) datasets. |
| |
|
| | ## Input Formatting |
| | Format input queries as follows: |
| | ``` |
| | input_text = f"Input:{input}\n\nResponse:" |
| | ``` |
| |
|
| | Note that due to an error during training, this model's end-of-sequence token ID is `0` instead of the `2` which is standard for Llama-2 based models. Inference APIs should handle this automatically by reading this repo's `config.json`, but be aware of this difference if you are doing token surgery. |
| |
|
| | ## Evals |
| | ` |
| | llemma_7b_muinstruct_camelmath` compares favorably to other 7B parameter models on the [Hungarian Math Exam](https://huggingface.co/datasets/keirp/hungarian_national_hs_finals_exam/blob/main/README.md). It surpasses the few-shot performance of Llemma 7B whilst being the strongest Llama-2 7B based model. |
| | |
| | | Model | Exam Score | |
| | | ------------------------------------------------------------------------------ | ---------- | |
| | | [Code Llama 7B](https://huggingface.co/codellama/CodeLlama-7b-hf) (few-shot) | 8\% | |
| | | [MetaMath 7B](https://huggingface.co/meta-math/MetaMath-7B-V1.0) | 20\% | |
| | | [MAmmoTH 7B](https://huggingface.co/TIGER-Lab/MAmmoTH-7B) | 17\% | |
| | | [MAmmoTH Coder 7B](https://huggingface.co/TIGER-Lab/MAmmoTH-Coder-7B) | 11\% | |
| | | [Llemma 7B](https://huggingface.co/EleutherAI/llemma_7b) (few-shot) | 23\% | |
| | | Llemma_7B_muinstruct_camelmath | 25\% | |
| | | - | - | |
| | | [Mistral 7B](https://huggingface.co/mistralai/Mistral-7B-v0.1) (few-shot) | 22\% | |
| | | [MetaMath Mistral 7B](https://huggingface.co/meta-math/MetaMath-Mistral-7B) | 29\% | |
| | | [OpenChat 3.5](https://huggingface.co/openchat/openchat_3.5) | 37\% | |
| |
|
| |
|
| |
|