Text Generation
Transformers
Safetensors
English
qwen2
qwq
reasoning
conversational
text-generation-inference
lbourdois commited on
Commit
c81d365
·
verified ·
1 Parent(s): 9a64478

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +101 -89
README.md CHANGED
@@ -1,90 +1,102 @@
1
- ---
2
- license: apache-2.0
3
- datasets:
4
- - gghfez/QwQ-LongCoT-130K-cleaned
5
- - qingy2024/QwQ-LongCoT-Verified-130K
6
- - amphora/QwQ-LongCoT-130K
7
- library_name: transformers
8
- language:
9
- - en
10
- base_model:
11
- - Qwen/Qwen2.5-0.5B-Instruct
12
- pipeline_tag: text-generation
13
- tags:
14
- - qwq
15
- - reasoning
16
- ---
17
-
18
- # **QWQ-500M [ Qwen Base]**
19
-
20
- QWQ-500M is a fine-tuned variant of Qwen2.5-0.5B, optimized for text generation tasks, particularly conversational reasoning and complex problem-solving. This model contains 494 million parameters and uses FP16 tensor type for efficient inference. It leverages the robust architecture of Qwen2.5 and has undergone further enhancements to excel in generating high-quality text, structured outputs, and multilingual support.
21
-
22
- ## **Key Features**
23
-
24
- 1. **Base Model**: Derived from [Qwen/Qwen2.5-0.5B](https://huggingface.co/Qwen/Qwen2.5-0.5B).
25
- 2. **Finetuned on Instruction Data**: Built upon Qwen2.5-0.5B-Instruct with specialized datasets for better instruction-following.
26
- 3. **Specialization**:
27
- - Advanced conversational reasoning.
28
- - Long-form content generation.
29
- - Support for generating structured data (JSON, tables).
30
- - Multilingual capabilities (over 29 languages).
31
- 4. **Optimized for Long Context**: Supports input contexts up to 128K tokens with generation capability up to 8K tokens.
32
-
33
- ---
34
-
35
- ## **Datasets Used**
36
-
37
- The model was fine-tuned on high-quality datasets explicitly curated for Chain of Thought (CoT) reasoning and long-context tasks. Notable datasets include:
38
-
39
- 1. **[amphora/QwQ-LongCoT-130K](https://huggingface.co/datasets/amphora/QwQ-LongCoT-130K)**: 133k samples focused on complex CoT reasoning.
40
- 2. **[qingy2024/QwQ-LongCoT-Verified-130K](https://huggingface.co/datasets/qingy2024/QwQ-LongCoT-Verified-130K)**: 467k verified samples emphasizing detailed step-by-step reasoning.
41
- 3. **[gghfez/QwQ-LongCoT-130K-cleaned](https://huggingface.co/datasets/gghfez/QwQ-LongCoT-130K-cleaned)**: 125k cleaned samples for high-accuracy reasoning tasks.
42
-
43
- ---
44
-
45
- ## **Running the Model**
46
-
47
- To run the model using the Transformers library:
48
-
49
- ```python
50
- # Install necessary libraries
51
- # pip install transformers torch
52
-
53
- from transformers import AutoTokenizer, AutoModelForCausalLM
54
- import torch
55
-
56
- tokenizer = AutoTokenizer.from_pretrained("prithivMLmods/QWQ-500M")
57
- model = AutoModelForCausalLM.from_pretrained(
58
- "prithivMLmods/QWQ-500M",
59
- torch_dtype=torch.float16,
60
- device_map="auto",
61
- )
62
-
63
- input_text = "Explain the concept of reinforcement learning."
64
- input_ids = tokenizer(input_text, return_tensors="pt").to("cuda")
65
-
66
- outputs = model.generate(**input_ids, max_new_tokens=100)
67
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))
68
- ```
69
-
70
- ---
71
-
72
- ## **Limitations**
73
-
74
- 1. **Bias and Fairness**: Despite fine-tuning efforts, biases from the training data may persist. Users should critically assess model outputs.
75
- 2. **Contextual Understanding**: While optimized for long contexts, the model may still occasionally misinterpret highly ambiguous prompts.
76
- 3. **Real-Time Knowledge**: The model's knowledge is limited to its training data and does not include real-time or post-training updates.
77
- 4. **Safety Considerations**: Safety alignment has been performed, but users should monitor outputs to avoid inappropriate content.
78
- 5. **Resource Requirements**: Running the model efficiently requires a GPU with sufficient memory.
79
-
80
- ---
81
-
82
- ## **Intended Use Cases**
83
-
84
- 1. **Conversational AI**: Enhanced dialogue capabilities with nuanced understanding and context retention.
85
- 2. **Educational Assistance**: Generating detailed explanations, tutorials, and step-by-step guides.
86
- 3. **Content Creation**: Assisting in writing blogs, articles, and creative content.
87
- 4. **Multilingual Applications**: Supporting content generation and translation across multiple languages.
88
- 5. **Data Generation**: Producing structured outputs such as JSON and tables for various applications.
89
-
 
 
 
 
 
 
 
 
 
 
 
 
90
  ---
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - gghfez/QwQ-LongCoT-130K-cleaned
5
+ - qingy2024/QwQ-LongCoT-Verified-130K
6
+ - amphora/QwQ-LongCoT-130K
7
+ library_name: transformers
8
+ language:
9
+ - zho
10
+ - eng
11
+ - fra
12
+ - spa
13
+ - por
14
+ - deu
15
+ - ita
16
+ - rus
17
+ - jpn
18
+ - kor
19
+ - vie
20
+ - tha
21
+ - ara
22
+ base_model:
23
+ - Qwen/Qwen2.5-0.5B-Instruct
24
+ pipeline_tag: text-generation
25
+ tags:
26
+ - qwq
27
+ - reasoning
28
+ ---
29
+
30
+ # **QWQ-500M [ Qwen Base]**
31
+
32
+ QWQ-500M is a fine-tuned variant of Qwen2.5-0.5B, optimized for text generation tasks, particularly conversational reasoning and complex problem-solving. This model contains 494 million parameters and uses FP16 tensor type for efficient inference. It leverages the robust architecture of Qwen2.5 and has undergone further enhancements to excel in generating high-quality text, structured outputs, and multilingual support.
33
+
34
+ ## **Key Features**
35
+
36
+ 1. **Base Model**: Derived from [Qwen/Qwen2.5-0.5B](https://huggingface.co/Qwen/Qwen2.5-0.5B).
37
+ 2. **Finetuned on Instruction Data**: Built upon Qwen2.5-0.5B-Instruct with specialized datasets for better instruction-following.
38
+ 3. **Specialization**:
39
+ - Advanced conversational reasoning.
40
+ - Long-form content generation.
41
+ - Support for generating structured data (JSON, tables).
42
+ - Multilingual capabilities (over 29 languages).
43
+ 4. **Optimized for Long Context**: Supports input contexts up to 128K tokens with generation capability up to 8K tokens.
44
+
45
+ ---
46
+
47
+ ## **Datasets Used**
48
+
49
+ The model was fine-tuned on high-quality datasets explicitly curated for Chain of Thought (CoT) reasoning and long-context tasks. Notable datasets include:
50
+
51
+ 1. **[amphora/QwQ-LongCoT-130K](https://huggingface.co/datasets/amphora/QwQ-LongCoT-130K)**: 133k samples focused on complex CoT reasoning.
52
+ 2. **[qingy2024/QwQ-LongCoT-Verified-130K](https://huggingface.co/datasets/qingy2024/QwQ-LongCoT-Verified-130K)**: 467k verified samples emphasizing detailed step-by-step reasoning.
53
+ 3. **[gghfez/QwQ-LongCoT-130K-cleaned](https://huggingface.co/datasets/gghfez/QwQ-LongCoT-130K-cleaned)**: 125k cleaned samples for high-accuracy reasoning tasks.
54
+
55
+ ---
56
+
57
+ ## **Running the Model**
58
+
59
+ To run the model using the Transformers library:
60
+
61
+ ```python
62
+ # Install necessary libraries
63
+ # pip install transformers torch
64
+
65
+ from transformers import AutoTokenizer, AutoModelForCausalLM
66
+ import torch
67
+
68
+ tokenizer = AutoTokenizer.from_pretrained("prithivMLmods/QWQ-500M")
69
+ model = AutoModelForCausalLM.from_pretrained(
70
+ "prithivMLmods/QWQ-500M",
71
+ torch_dtype=torch.float16,
72
+ device_map="auto",
73
+ )
74
+
75
+ input_text = "Explain the concept of reinforcement learning."
76
+ input_ids = tokenizer(input_text, return_tensors="pt").to("cuda")
77
+
78
+ outputs = model.generate(**input_ids, max_new_tokens=100)
79
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
80
+ ```
81
+
82
+ ---
83
+
84
+ ## **Limitations**
85
+
86
+ 1. **Bias and Fairness**: Despite fine-tuning efforts, biases from the training data may persist. Users should critically assess model outputs.
87
+ 2. **Contextual Understanding**: While optimized for long contexts, the model may still occasionally misinterpret highly ambiguous prompts.
88
+ 3. **Real-Time Knowledge**: The model's knowledge is limited to its training data and does not include real-time or post-training updates.
89
+ 4. **Safety Considerations**: Safety alignment has been performed, but users should monitor outputs to avoid inappropriate content.
90
+ 5. **Resource Requirements**: Running the model efficiently requires a GPU with sufficient memory.
91
+
92
+ ---
93
+
94
+ ## **Intended Use Cases**
95
+
96
+ 1. **Conversational AI**: Enhanced dialogue capabilities with nuanced understanding and context retention.
97
+ 2. **Educational Assistance**: Generating detailed explanations, tutorials, and step-by-step guides.
98
+ 3. **Content Creation**: Assisting in writing blogs, articles, and creative content.
99
+ 4. **Multilingual Applications**: Supporting content generation and translation across multiple languages.
100
+ 5. **Data Generation**: Producing structured outputs such as JSON and tables for various applications.
101
+
102
  ---