CannaeAI commited on
Commit
3031856
·
verified ·
1 Parent(s): e460ff2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -24,13 +24,13 @@ This is a fine-tuned version of [unsloth/Llama-3.2-1B](https://huggingface.co/un
24
  - **License :** apache-2.0
25
  - **Finetuned from model :** unsloth/Llama-3.2-1B
26
  ## Benchmarks:
27
- We evaluate ReasoningLlama-Math-1B-IT on GSM8K using the standard lm-eval 5-shot exact-match protocol. Under identical decoding and extraction settings, the model achieves:
28
  | Model | Params | GSM8K (5-shot, EM) |
29
  | ----------------------------- | ------ | ------------------ |
30
  | **ReasoningLlama-Math-1B-IT** | 1B | **30.7%** |
31
- | Llama-3.2-1B-Instruct | 1B | 6.8% |
32
 
33
 
34
  <p align="center">
35
- <img alt="gpt-oss-safeguard-20b" src="https://huggingface.co/Cannae-AI/ReasoningLlama-Math-1B-IT/resolve/main/ChatGPT%20Image%2018%20nov.%202025%2C%2020_55_23.png">
36
  </p>
 
24
  - **License :** apache-2.0
25
  - **Finetuned from model :** unsloth/Llama-3.2-1B
26
  ## Benchmarks:
27
+ We evaluate ReasoningLlama-Math-1B-IT on GSM8K using the standard lm-eval 5-shot exact-match protocol. Under identical decoding and extraction settings,ReasoningLlama-Math-1B-IT outperforms Meta’s Llama-3.2-1B by more than 4×, demonstrating a substantial improvement in small-model mathematical reasoning capability.
28
  | Model | Params | GSM8K (5-shot, EM) |
29
  | ----------------------------- | ------ | ------------------ |
30
  | **ReasoningLlama-Math-1B-IT** | 1B | **30.7%** |
31
+ | Llama-3.2-1B | 1B | 6.8% |
32
 
33
 
34
  <p align="center">
35
+ <img alt="ReasonningLlama-Math-1B-IT" src="https://huggingface.co/Cannae-AI/ReasoningLlama-Math-1B-IT/resolve/main/ChatGPT%20Image%2018%20nov.%202025%2C%2020_55_23.png">
36
  </p>