Text Generation
Transformers
Safetensors
English

Add pipeline tag and update library_name

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -1,15 +1,16 @@
1
  ---
2
  base_model: unsloth/qwen2.5-1.5b-instruct-unsloth-bnb-4bit
3
- library_name: peft
4
- license: apache-2.0
5
  datasets:
6
  - openai/gsm8k
7
  - HuggingFaceH4/MATH-500
8
  - HuggingFaceH4/aime_2024
9
  language:
10
  - en
 
 
11
  metrics:
12
  - accuracy
 
13
  ---
14
 
15
  ## MOTIF: Modular Thinking via Reinforcement Fine-tuning in LLMs
@@ -49,7 +50,7 @@ from transformers import AutoModelForCausalLM
49
  base_model = AutoModelForCausalLM.from_pretrained("unsloth/qwen2.5-1.5b-instruct-unsloth-bnb-4bit")
50
  model = PeftModel.from_pretrained(base_model, "purbeshmitra/vanillaGRPO")
51
 
52
- SYSTEM_PROMPT = "You are a helpful assistant. When the user asks a question, you first think about the reasoning process in mind and then provide the user with an answer. The reasoning process and the answer are enclosed within <reasoning> </reasoning> and <answer> </answer> tags, respectively. In your answer, you also enclose your final answer in the box: \boxed{}. Therefore, you respond in the following strict format:
53
  <reasoning> reasoning process here </reasoning> <answer> answer here </answer>."
54
  ```
55
 
 
1
  ---
2
  base_model: unsloth/qwen2.5-1.5b-instruct-unsloth-bnb-4bit
 
 
3
  datasets:
4
  - openai/gsm8k
5
  - HuggingFaceH4/MATH-500
6
  - HuggingFaceH4/aime_2024
7
  language:
8
  - en
9
+ library_name: transformers
10
+ license: apache-2.0
11
  metrics:
12
  - accuracy
13
+ pipeline_tag: text-generation
14
  ---
15
 
16
  ## MOTIF: Modular Thinking via Reinforcement Fine-tuning in LLMs
 
50
  base_model = AutoModelForCausalLM.from_pretrained("unsloth/qwen2.5-1.5b-instruct-unsloth-bnb-4bit")
51
  model = PeftModel.from_pretrained(base_model, "purbeshmitra/vanillaGRPO")
52
 
53
+ SYSTEM_PROMPT = "You are a helpful assistant. When the user asks a question, you first think about the reasoning process in mind and then provide the user with an answer. The reasoning process and the answer are enclosed within <reasoning> </reasoning> and <answer> </answer> tags, respectively. In your answer, you also enclose your final answer in the box: \\boxed{}. Therefore, you respond in the following strict format:
54
  <reasoning> reasoning process here </reasoning> <answer> answer here </answer>."
55
  ```
56