Improve language tag

#1
by lbourdois - opened
Files changed (1) hide show
  1. README.md +92 -80
README.md CHANGED
@@ -1,81 +1,93 @@
1
- ---
2
- license: apache-2.0
3
- datasets:
4
- - openai/gsm8k
5
- language:
6
- - en
7
- base_model:
8
- - Qwen/Qwen2.5-0.5B-Instruct
9
- pipeline_tag: text-generation
10
- library_name: transformers
11
- tags:
12
- - math
13
- - reasoning
14
- - grpo
15
- - trl
16
- - code
17
- ---
18
- # **Feynman-Grpo-Exp**
19
-
20
- Feynman-Grpo-Exp is based on the Qwen 0.5B modality architecture, designed to enhance the reasoning capabilities of 0.5B-parameter models. It has been fine-tuned using the GRPO trainer on the OpenAI GSM8K dataset for reinforcement learning, improving its ability to handle complex reasoning tasks, multi-step problem-solving, and mathematical challenges. This model excels in chain-of-thought (CoT) reasoning and logical problem-solving, making it suitable for a variety of advanced tasks that require precise and structured outputs.
21
-
22
- ### **Key Improvements**
23
- 1. **Enhanced Knowledge and Expertise**: Strengthened mathematical reasoning, code generation, and problem-solving skills, particularly in scientific and technical domains.
24
- 2. **Fine-Tuned Instruction Following**: Optimized for generating structured outputs like JSON and handling long-form text (up to 8K+ tokens).
25
- 3. **Greater Adaptability**: Enhanced role-playing capabilities, allowing for better responses to diverse prompts.
26
- 4. **Long-Context Support**: Capable of processing up to **64K tokens** and generating up to **4K tokens** per output.
27
- 5. **Multilingual Proficiency**: Supports over **29 languages**, including Chinese, English, French, Spanish, Portuguese, German, and more.
28
-
29
- ### **Quickstart with Transformers**
30
-
31
- ```python
32
- from transformers import AutoModelForCausalLM, AutoTokenizer
33
-
34
- model_name = "prithivMLmods/Feynman-Grpo-Exp"
35
-
36
- model = AutoModelForCausalLM.from_pretrained(
37
- model_name,
38
- torch_dtype="auto",
39
- device_map="auto",
40
- trust_remote_code=True
41
- )
42
- tokenizer = AutoTokenizer.from_pretrained(model_name)
43
-
44
- prompt = "Give me a short introduction to large language models."
45
- messages = [
46
- {"role": "system", "content": "You are an advanced AI assistant with expert-level reasoning and knowledge."},
47
- {"role": "user", "content": prompt}
48
- ]
49
- text = tokenizer.apply_chat_template(
50
- messages,
51
- tokenize=False,
52
- add_generation_prompt=True
53
- )
54
- model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
55
-
56
- generated_ids = model.generate(
57
- **model_inputs,
58
- max_new_tokens=512
59
- )
60
- generated_ids = [
61
- output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
62
- ]
63
-
64
- response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
65
- print(response)
66
- ```
67
-
68
- ### **Intended Use**
69
- - **Advanced Reasoning & Context Understanding**: Ideal for logical deduction, multi-step problem-solving, and complex knowledge-based tasks.
70
- - **Mathematical & Scientific Problem-Solving**: Optimized for handling advanced calculations, theorem proving, and scientific queries.
71
- - **Code Generation & Debugging**: Capable of generating and optimizing code across multiple programming languages.
72
- - **Structured Data Analysis**: Processes structured data, including tables, JSON, and other formats, making it well-suited for data-centric tasks.
73
- - **Multilingual Applications**: Proficient in over 29 languages, enabling a global scale for applications.
74
- - **Extended Content Generation**: Supports detailed document writing, research reports, and instructional guides.
75
-
76
- ### **Limitations**
77
- 1. **Computational Requirements**: Despite being a **0.5B-parameter** model, it requires significant computational resources for efficient inference, especially when dealing with long-context processing.
78
- 2. **Language-Specific Variability**: Performance may vary across supported languages, with possible challenges for low-resource languages.
79
- 3. **Potential Error Accumulation**: Long-text generation can introduce inconsistencies or errors over extended outputs.
80
- 4. **Limited Real-World Awareness**: The model's knowledge is restricted to the training data, which may not reflect the most recent events or developments.
 
 
 
 
 
 
 
 
 
 
 
 
81
  5. **Prompt Sensitivity**: Outputs depend heavily on the specificity and clarity of the input prompts.
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - openai/gsm8k
5
+ language:
6
+ - zho
7
+ - eng
8
+ - fra
9
+ - spa
10
+ - por
11
+ - deu
12
+ - ita
13
+ - rus
14
+ - jpn
15
+ - kor
16
+ - vie
17
+ - tha
18
+ - ara
19
+ base_model:
20
+ - Qwen/Qwen2.5-0.5B-Instruct
21
+ pipeline_tag: text-generation
22
+ library_name: transformers
23
+ tags:
24
+ - math
25
+ - reasoning
26
+ - grpo
27
+ - trl
28
+ - code
29
+ ---
30
+ # **Feynman-Grpo-Exp**
31
+
32
+ Feynman-Grpo-Exp is based on the Qwen 0.5B modality architecture, designed to enhance the reasoning capabilities of 0.5B-parameter models. It has been fine-tuned using the GRPO trainer on the OpenAI GSM8K dataset for reinforcement learning, improving its ability to handle complex reasoning tasks, multi-step problem-solving, and mathematical challenges. This model excels in chain-of-thought (CoT) reasoning and logical problem-solving, making it suitable for a variety of advanced tasks that require precise and structured outputs.
33
+
34
+ ### **Key Improvements**
35
+ 1. **Enhanced Knowledge and Expertise**: Strengthened mathematical reasoning, code generation, and problem-solving skills, particularly in scientific and technical domains.
36
+ 2. **Fine-Tuned Instruction Following**: Optimized for generating structured outputs like JSON and handling long-form text (up to 8K+ tokens).
37
+ 3. **Greater Adaptability**: Enhanced role-playing capabilities, allowing for better responses to diverse prompts.
38
+ 4. **Long-Context Support**: Capable of processing up to **64K tokens** and generating up to **4K tokens** per output.
39
+ 5. **Multilingual Proficiency**: Supports over **29 languages**, including Chinese, English, French, Spanish, Portuguese, German, and more.
40
+
41
+ ### **Quickstart with Transformers**
42
+
43
+ ```python
44
+ from transformers import AutoModelForCausalLM, AutoTokenizer
45
+
46
+ model_name = "prithivMLmods/Feynman-Grpo-Exp"
47
+
48
+ model = AutoModelForCausalLM.from_pretrained(
49
+ model_name,
50
+ torch_dtype="auto",
51
+ device_map="auto",
52
+ trust_remote_code=True
53
+ )
54
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
55
+
56
+ prompt = "Give me a short introduction to large language models."
57
+ messages = [
58
+ {"role": "system", "content": "You are an advanced AI assistant with expert-level reasoning and knowledge."},
59
+ {"role": "user", "content": prompt}
60
+ ]
61
+ text = tokenizer.apply_chat_template(
62
+ messages,
63
+ tokenize=False,
64
+ add_generation_prompt=True
65
+ )
66
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
67
+
68
+ generated_ids = model.generate(
69
+ **model_inputs,
70
+ max_new_tokens=512
71
+ )
72
+ generated_ids = [
73
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
74
+ ]
75
+
76
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
77
+ print(response)
78
+ ```
79
+
80
+ ### **Intended Use**
81
+ - **Advanced Reasoning & Context Understanding**: Ideal for logical deduction, multi-step problem-solving, and complex knowledge-based tasks.
82
+ - **Mathematical & Scientific Problem-Solving**: Optimized for handling advanced calculations, theorem proving, and scientific queries.
83
+ - **Code Generation & Debugging**: Capable of generating and optimizing code across multiple programming languages.
84
+ - **Structured Data Analysis**: Processes structured data, including tables, JSON, and other formats, making it well-suited for data-centric tasks.
85
+ - **Multilingual Applications**: Proficient in over 29 languages, enabling a global scale for applications.
86
+ - **Extended Content Generation**: Supports detailed document writing, research reports, and instructional guides.
87
+
88
+ ### **Limitations**
89
+ 1. **Computational Requirements**: Despite being a **0.5B-parameter** model, it requires significant computational resources for efficient inference, especially when dealing with long-context processing.
90
+ 2. **Language-Specific Variability**: Performance may vary across supported languages, with possible challenges for low-resource languages.
91
+ 3. **Potential Error Accumulation**: Long-text generation can introduce inconsistencies or errors over extended outputs.
92
+ 4. **Limited Real-World Awareness**: The model's knowledge is restricted to the training data, which may not reflect the most recent events or developments.
93
  5. **Prompt Sensitivity**: Outputs depend heavily on the specificity and clarity of the input prompts.