Improve language tag

#1
by lbourdois - opened
Files changed (1) hide show
  1. README.md +117 -106
README.md CHANGED
@@ -1,107 +1,118 @@
1
- ---
2
- library_name: transformers
3
- base_model:
4
- - Qwen/Qwen2.5-3B-Instruct
5
- license: apache-2.0
6
- datasets:
7
- - amphora/QwQ-LongCoT-130K
8
- - amphora/QwQ-LongCoT-130K-2
9
- - amphora/verfiable-25k
10
- - amphora/m-math500
11
- language:
12
- - en
13
- - zh
14
- pipeline_tag: text-generation
15
- tags:
16
- - Math
17
- - Code
18
- - Thinker
19
- - Reasoning
20
- - 3B
21
- - QwQ
22
- - Mini
23
- - text-generation-inference
24
- - SFT
25
- - trl
26
- ---
27
-
28
- ![8.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/Faj50x1HAODJAIy_R94se.png)
29
-
30
- # **PocketThinker-QwQ-3B-Instruct**
31
-
32
- > PocketThinker-QwQ-3B-Instruct is based on the Qwen2.5-3B-Instruct architecture, designed as a lightweight and efficient reasoning assistant. It serves as the pocket-sized version of QwQ-LCoT-7B-Instruct, optimized for fast inference while maintaining strong problem-solving and computational capabilities. This model is fine-tuned for enhanced structured reasoning, minimal token wastage, and high-quality technical responses.
33
-
34
- ## **Key Improvements**
35
- 1. **Optimized for Coding**: Specializes in generating structured, efficient code with minimal redundancy for smooth execution.
36
- 2. **Compact yet Powerful**: Maintains strong problem-solving capabilities within a smaller 3B parameter architecture, ensuring accessibility on resource-limited devices.
37
- 3. **Advanced Reasoning Capabilities**: Excels in algorithmic problem-solving, mathematical reasoning, and structured technical explanations.
38
- 4. **Efficient Memory Utilization**: Reduces computational overhead while maintaining high-quality outputs.
39
- 5. **Focused Output Generation**: Avoids unnecessary token generation, ensuring concise and relevant responses.
40
-
41
- ## **Quickstart with transformers**
42
-
43
- Here is a code snippet to load the tokenizer and model using `apply_chat_template` for structured input formatting:
44
-
45
- ```python
46
- from transformers import AutoModelForCausalLM, AutoTokenizer
47
-
48
- model_name = "prithivMLmods/PocketThinker-QwQ-3B-Instruct"
49
-
50
- model = AutoModelForCausalLM.from_pretrained(
51
- model_name,
52
- torch_dtype="auto",
53
- device_map="auto"
54
- )
55
- tokenizer = AutoTokenizer.from_pretrained(model_name)
56
-
57
- prompt = "Write a Python function to find the Fibonacci sequence."
58
- messages = [
59
- {"role": "system", "content": "You are an advanced coding assistant."},
60
- {"role": "user", "content": prompt}
61
- ]
62
- text = tokenizer.apply_chat_template(
63
- messages,
64
- tokenize=False,
65
- add_generation_prompt=True
66
- )
67
- model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
68
-
69
- generated_ids = model.generate(
70
- **model_inputs,
71
- max_new_tokens=6090
72
- )
73
- generated_ids = [
74
- output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
75
- ]
76
-
77
- response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
78
- print(response)
79
- ```
80
-
81
- ## **Intended Use**
82
- 1. **Code Generation & Optimization**:
83
- Supports developers in writing, refining, and optimizing code across multiple programming languages.
84
- 2. **Algorithm & Mathematical Problem Solving**:
85
- Delivers precise solutions and structured explanations for complex problems.
86
- 3. **Technical Documentation & Explanation**:
87
- Assists in generating well-structured documentation for libraries, APIs, and coding concepts.
88
- 4. **Debugging Assistance**:
89
- Helps identify and correct errors in code snippets.
90
- 5. **Educational Support**:
91
- Simplifies programming topics for students and learners with clear explanations.
92
- 6. **Structured Data Processing**:
93
- Generates structured outputs like JSON, XML, and tables for data science applications.
94
-
95
- ## **Limitations**
96
- 1. **Hardware Constraints**:
97
- Although lighter than larger models, still requires a moderately powerful GPU or TPU for optimal performance.
98
- 2. **Potential Bias in Responses**:
99
- Outputs may reflect biases present in training data.
100
- 3. **Limited Creativity**:
101
- May generate variable results in non-technical, creative tasks.
102
- 4. **No Real-Time Awareness**:
103
- Lacks access to real-world events beyond its training cutoff.
104
- 5. **Error Propagation in Long Responses**:
105
- Minor mistakes in early outputs may affect overall coherence in lengthy responses.
106
- 6. **Prompt Sensitivity**:
 
 
 
 
 
 
 
 
 
 
 
107
  The effectiveness of responses depends on well-structured prompts.
 
1
+ ---
2
+ library_name: transformers
3
+ base_model:
4
+ - Qwen/Qwen2.5-3B-Instruct
5
+ license: apache-2.0
6
+ datasets:
7
+ - amphora/QwQ-LongCoT-130K
8
+ - amphora/QwQ-LongCoT-130K-2
9
+ - amphora/verfiable-25k
10
+ - amphora/m-math500
11
+ language:
12
+ - zho
13
+ - eng
14
+ - fra
15
+ - spa
16
+ - por
17
+ - deu
18
+ - ita
19
+ - rus
20
+ - jpn
21
+ - kor
22
+ - vie
23
+ - tha
24
+ - ara
25
+ pipeline_tag: text-generation
26
+ tags:
27
+ - Math
28
+ - Code
29
+ - Thinker
30
+ - Reasoning
31
+ - 3B
32
+ - QwQ
33
+ - Mini
34
+ - text-generation-inference
35
+ - SFT
36
+ - trl
37
+ ---
38
+
39
+ ![8.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/Faj50x1HAODJAIy_R94se.png)
40
+
41
+ # **PocketThinker-QwQ-3B-Instruct**
42
+
43
+ > PocketThinker-QwQ-3B-Instruct is based on the Qwen2.5-3B-Instruct architecture, designed as a lightweight and efficient reasoning assistant. It serves as the pocket-sized version of QwQ-LCoT-7B-Instruct, optimized for fast inference while maintaining strong problem-solving and computational capabilities. This model is fine-tuned for enhanced structured reasoning, minimal token wastage, and high-quality technical responses.
44
+
45
+ ## **Key Improvements**
46
+ 1. **Optimized for Coding**: Specializes in generating structured, efficient code with minimal redundancy for smooth execution.
47
+ 2. **Compact yet Powerful**: Maintains strong problem-solving capabilities within a smaller 3B parameter architecture, ensuring accessibility on resource-limited devices.
48
+ 3. **Advanced Reasoning Capabilities**: Excels in algorithmic problem-solving, mathematical reasoning, and structured technical explanations.
49
+ 4. **Efficient Memory Utilization**: Reduces computational overhead while maintaining high-quality outputs.
50
+ 5. **Focused Output Generation**: Avoids unnecessary token generation, ensuring concise and relevant responses.
51
+
52
+ ## **Quickstart with transformers**
53
+
54
+ Here is a code snippet to load the tokenizer and model using `apply_chat_template` for structured input formatting:
55
+
56
+ ```python
57
+ from transformers import AutoModelForCausalLM, AutoTokenizer
58
+
59
+ model_name = "prithivMLmods/PocketThinker-QwQ-3B-Instruct"
60
+
61
+ model = AutoModelForCausalLM.from_pretrained(
62
+ model_name,
63
+ torch_dtype="auto",
64
+ device_map="auto"
65
+ )
66
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
67
+
68
+ prompt = "Write a Python function to find the Fibonacci sequence."
69
+ messages = [
70
+ {"role": "system", "content": "You are an advanced coding assistant."},
71
+ {"role": "user", "content": prompt}
72
+ ]
73
+ text = tokenizer.apply_chat_template(
74
+ messages,
75
+ tokenize=False,
76
+ add_generation_prompt=True
77
+ )
78
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
79
+
80
+ generated_ids = model.generate(
81
+ **model_inputs,
82
+ max_new_tokens=6090
83
+ )
84
+ generated_ids = [
85
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
86
+ ]
87
+
88
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
89
+ print(response)
90
+ ```
91
+
92
+ ## **Intended Use**
93
+ 1. **Code Generation & Optimization**:
94
+ Supports developers in writing, refining, and optimizing code across multiple programming languages.
95
+ 2. **Algorithm & Mathematical Problem Solving**:
96
+ Delivers precise solutions and structured explanations for complex problems.
97
+ 3. **Technical Documentation & Explanation**:
98
+ Assists in generating well-structured documentation for libraries, APIs, and coding concepts.
99
+ 4. **Debugging Assistance**:
100
+ Helps identify and correct errors in code snippets.
101
+ 5. **Educational Support**:
102
+ Simplifies programming topics for students and learners with clear explanations.
103
+ 6. **Structured Data Processing**:
104
+ Generates structured outputs like JSON, XML, and tables for data science applications.
105
+
106
+ ## **Limitations**
107
+ 1. **Hardware Constraints**:
108
+ Although lighter than larger models, still requires a moderately powerful GPU or TPU for optimal performance.
109
+ 2. **Potential Bias in Responses**:
110
+ Outputs may reflect biases present in training data.
111
+ 3. **Limited Creativity**:
112
+ May generate variable results in non-technical, creative tasks.
113
+ 4. **No Real-Time Awareness**:
114
+ Lacks access to real-world events beyond its training cutoff.
115
+ 5. **Error Propagation in Long Responses**:
116
+ Minor mistakes in early outputs may affect overall coherence in lengthy responses.
117
+ 6. **Prompt Sensitivity**:
118
  The effectiveness of responses depends on well-structured prompts.