Add metadata and link to the paper and code

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +20 -7
README.md CHANGED
@@ -1,13 +1,20 @@
 
 
 
 
 
1
  ## Introduction
2
  MetaStone-L1 is the lite reasoning model of the MetaStone series, which aims to enhance the performance in hard downstream tasks.
3
 
4
  On core reasoning benchmarks including mathematics and code, MetaStone-L1-7B achieved SOTA results in the parallel-level models, and it also achieved the comparable results as the API models such as Claude-3.5-Sonnet-1022 and GPT4o-0513.
5
  <img src="./introduction.png" alt="Logo" width="800">
6
 
7
- This repo contains the MetaStone-L1-7B model, which is trained based on DeepSeek-R1-Distill-Qwen-7B by GRPO. For full details of this model please refer to our release blog.
 
 
8
 
9
  ## Requirements
10
- We advise you to use the latest version of transformers(```transformers==4.48.3```). For the best experience, please review the [Usage Guidelines](#usage-guidelines).
11
 
12
  ## Quickstart
13
  Here give the example of how to use our model.
@@ -20,7 +27,9 @@ tokenizer = AutoTokenizer.from_pretrained(model_name)
20
  model = AutoModelForCausalLM.from_pretrained(model_name)
21
 
22
  messages = [
23
- {"role": "user", "content": "Complete the square for the following quadratic: $-x^2+7 x-11$\n\nPlease reason step by step, and put your final answer within \\boxed{}."}
 
 
24
  ]
25
 
26
  text = tokenizer.apply_chat_template(
@@ -46,7 +55,8 @@ To achieve optimal performance, we recommend the following settings:
46
 
47
  1. Enhace the thoughful output:
48
 
49
- a. Make sure the model starts with ```<think>\n``` to prevent generating empty think content. If you use ```apply_chat_template``` and set ```add_generation_prompt=True```, this is automatically implemented, but this may result in replies not having a <think> tag at the beginning, which is normal.
 
50
 
51
  b. Ensure the final input of the model is in the format of ```<|User|> [your prompt] <|Assistant|><think>```.
52
 
@@ -56,7 +66,11 @@ To achieve optimal performance, we recommend the following settings:
56
 
57
  a. Math questions: Add a statement "```Please reason step by step, and put your final answer within \\boxed{}.```" to the prompt.
58
 
59
- b. Code problems: Add "### Format: Read the inputs from stdin solve the problem and write the answer to stdout. Enclose your code within delimiters as follows.\n \```python\n# YOUR CODE HERE\n\```\n### Answer: (use the provided format with backticks)" to the prompt.
 
 
 
 
60
 
61
  4. In particular, we use ```latex2sympy2``` and ```sympy``` to assist in judging complex Latex formats for the Math500 evaluation script. For all datasets, we generate 64 responses per query to estimate pass@1.
62
 
@@ -79,5 +93,4 @@ If you find our work helpful, feel free to give us a cite.
79
  journal={arXiv preprint arXiv:2412.08864},
80
  year={2024}
81
  }
82
- ```
83
-
 
1
+ ---
2
+ pipeline_tag: text-generation
3
+ library_name: transformers
4
+ ---
5
+
6
  ## Introduction
7
  MetaStone-L1 is the lite reasoning model of the MetaStone series, which aims to enhance the performance in hard downstream tasks.
8
 
9
  On core reasoning benchmarks including mathematics and code, MetaStone-L1-7B achieved SOTA results in the parallel-level models, and it also achieved the comparable results as the API models such as Claude-3.5-Sonnet-1022 and GPT4o-0513.
10
  <img src="./introduction.png" alt="Logo" width="800">
11
 
12
+ This repo contains the MetaStone-L1-7B model, which is trained based on DeepSeek-R1-Distill-Qwen-7B by GRPO. For full details of this model please refer to our release blog. The model was presented in [A Graph-Based Synthetic Data Pipeline for Scaling High-Quality Reasoning Instructions](https://huggingface.co/papers/2412.08864).
13
+
14
+ Code: https://github.com/Jayce1kk/GSDP
15
 
16
  ## Requirements
17
+ We advise you to use the latest version of transformers(```transformers==4.48.3```). For the best experience, please review the [Usage Guidelines](#usage-guidelines).
18
 
19
  ## Quickstart
20
  Here give the example of how to use our model.
 
27
  model = AutoModelForCausalLM.from_pretrained(model_name)
28
 
29
  messages = [
30
+ {"role": "user", "content": "Complete the square for the following quadratic: $-x^2+7 x-11$
31
+
32
+ Please reason step by step, and put your final answer within \\boxed{}."}
33
  ]
34
 
35
  text = tokenizer.apply_chat_template(
 
55
 
56
  1. Enhace the thoughful output:
57
 
58
+ a. Make sure the model starts with ```<think>
59
+ ``` to prevent generating empty think content. If you use ```apply_chat_template``` and set ```add_generation_prompt=True```, this is automatically implemented, but this may result in replies not having a <think> tag at the beginning, which is normal.
60
 
61
  b. Ensure the final input of the model is in the format of ```<|User|> [your prompt] <|Assistant|><think>```.
62
 
 
66
 
67
  a. Math questions: Add a statement "```Please reason step by step, and put your final answer within \\boxed{}.```" to the prompt.
68
 
69
+ b. Code problems: Add "### Format: Read the inputs from stdin solve the problem and write the answer to stdout. Enclose your code within delimiters as follows.
70
+ \```python
71
+ # YOUR CODE HERE
72
+ \```
73
+ ### Answer: (use the provided format with backticks)" to the prompt.
74
 
75
  4. In particular, we use ```latex2sympy2``` and ```sympy``` to assist in judging complex Latex formats for the Math500 evaluation script. For all datasets, we generate 64 responses per query to estimate pass@1.
76
 
 
93
  journal={arXiv preprint arXiv:2412.08864},
94
  year={2024}
95
  }
96
+ ```