JingyaoLi commited on
Commit
e462a06
·
verified ·
1 Parent(s): 62be290

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +15 -10
README.md CHANGED
@@ -15,21 +15,26 @@ model-index:
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
  # ScienceLLaMA-3B
18
- This model is a fine-tuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) on the OpenMathInstruct-2-1M and the metamath_gsm8k datasets.
 
 
 
 
 
 
19
 
20
- ## Model description
21
 
22
- More information needed
 
 
23
 
24
- ## Intended uses & limitations
25
-
26
- More information needed
27
-
28
- ## Training and evaluation data
29
-
30
- More information needed
31
 
32
  ## Training procedure
 
 
 
 
33
 
34
  ### Training hyperparameters
35
 
 
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
  # ScienceLLaMA-3B
18
+ <p align="center">
19
+ • 🤗 <a href="https://huggingface.co/datasets/JingyaoLi/Science-Logits-1.2M" target="_blank">Data </a>
20
+ • 🤗 <a href="https://huggingface.co/JingyaoLi/ScienceLLaMA-3b" target="_blank">ScienceLLaMA-3B </a>
21
+ • 🤗 <a href="https://huggingface.co/JingyaoLi/ScienceLLaMA-1b" target="_blank">ScienceLLaMA-1B </a>
22
+ • 🐱 <a href="Logits-based Finetuning" target="_blank">Code</a>
23
+ • 📃 Paper (TO be released) <br>
24
+ </p>
25
 
26
+ This model is a fine-tuned with **Logits-Based Finetuning** on the [JingyaoLi/Science-Logits-1.2M](https://huggingface.co/datasets/JingyaoLi/Science-Logits-1.2M), which integrates the strengths of supervised learning and knowledge distillation by combining teacher logits with ground truth labels. This preserves both correctness and linguistic diversity.
27
 
28
+ <div style="text-align: center;">
29
+ <img src="./images/example.png" alt="example" />
30
+ </div>
31
 
 
 
 
 
 
 
 
32
 
33
  ## Training procedure
34
+ <div style="text-align: center;">
35
+ <img src="./images/performance.png" alt="performance" />
36
+ </div>
37
+
38
 
39
  ### Training hyperparameters
40