Improve language tag

#1
by lbourdois - opened
Files changed (1) hide show
  1. README.md +63 -50
README.md CHANGED
@@ -1,50 +1,63 @@
1
- ---
2
- base_model:
3
- - Qwen/Qwen2.5-3B-Instruct
4
- - bunnycore/Qwen-2.5-3b-Rp-lora_model
5
- - bunnycore/Qwen-2.5-3b-Evol-CoT
6
- - bunnycore/Qwen-2.5-3b-Rp-lora_model
7
- - Qwen/Qwen2.5-3B
8
- library_name: transformers
9
- tags:
10
- - mergekit
11
- - merge
12
-
13
- ---
14
- # merge
15
-
16
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
17
-
18
- ## Merge Details
19
- ### Merge Method
20
-
21
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [Qwen/Qwen2.5-3B](https://huggingface.co/Qwen/Qwen2.5-3B) as a base.
22
-
23
- ### Models Merged
24
-
25
- The following models were included in the merge:
26
- * [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct) + [bunnycore/Qwen-2.5-3b-Rp-lora_model](https://huggingface.co/bunnycore/Qwen-2.5-3b-Rp-lora_model)
27
- * [bunnycore/Qwen-2.5-3b-Evol-CoT](https://huggingface.co/bunnycore/Qwen-2.5-3b-Evol-CoT) + [bunnycore/Qwen-2.5-3b-Rp-lora_model](https://huggingface.co/bunnycore/Qwen-2.5-3b-Rp-lora_model)
28
-
29
- ### Configuration
30
-
31
- The following YAML configuration was used to produce this model:
32
-
33
- ```yaml
34
- models:
35
- - model: Qwen/Qwen2.5-3B-Instruct+bunnycore/Qwen-2.5-3b-Rp-lora_model
36
- parameters:
37
- density: 0.5
38
- weight: 0.5
39
- - model: bunnycore/Qwen-2.5-3b-Evol-CoT+bunnycore/Qwen-2.5-3b-Rp-lora_model
40
- parameters:
41
- density: 0.5
42
- weight: 0.5
43
-
44
- merge_method: ties
45
- base_model: Qwen/Qwen2.5-3B
46
- parameters:
47
- normalize: false
48
- int8_mask: true
49
- dtype: float16
50
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-3B-Instruct
4
+ - bunnycore/Qwen-2.5-3b-Rp-lora_model
5
+ - bunnycore/Qwen-2.5-3b-Evol-CoT
6
+ - bunnycore/Qwen-2.5-3b-Rp-lora_model
7
+ - Qwen/Qwen2.5-3B
8
+ library_name: transformers
9
+ tags:
10
+ - mergekit
11
+ - merge
12
+ language:
13
+ - zho
14
+ - eng
15
+ - fra
16
+ - spa
17
+ - por
18
+ - deu
19
+ - ita
20
+ - rus
21
+ - jpn
22
+ - kor
23
+ - vie
24
+ - tha
25
+ - ara
26
+ ---
27
+ # merge
28
+
29
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
30
+
31
+ ## Merge Details
32
+ ### Merge Method
33
+
34
+ This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [Qwen/Qwen2.5-3B](https://huggingface.co/Qwen/Qwen2.5-3B) as a base.
35
+
36
+ ### Models Merged
37
+
38
+ The following models were included in the merge:
39
+ * [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct) + [bunnycore/Qwen-2.5-3b-Rp-lora_model](https://huggingface.co/bunnycore/Qwen-2.5-3b-Rp-lora_model)
40
+ * [bunnycore/Qwen-2.5-3b-Evol-CoT](https://huggingface.co/bunnycore/Qwen-2.5-3b-Evol-CoT) + [bunnycore/Qwen-2.5-3b-Rp-lora_model](https://huggingface.co/bunnycore/Qwen-2.5-3b-Rp-lora_model)
41
+
42
+ ### Configuration
43
+
44
+ The following YAML configuration was used to produce this model:
45
+
46
+ ```yaml
47
+ models:
48
+ - model: Qwen/Qwen2.5-3B-Instruct+bunnycore/Qwen-2.5-3b-Rp-lora_model
49
+ parameters:
50
+ density: 0.5
51
+ weight: 0.5
52
+ - model: bunnycore/Qwen-2.5-3b-Evol-CoT+bunnycore/Qwen-2.5-3b-Rp-lora_model
53
+ parameters:
54
+ density: 0.5
55
+ weight: 0.5
56
+
57
+ merge_method: ties
58
+ base_model: Qwen/Qwen2.5-3B
59
+ parameters:
60
+ normalize: false
61
+ int8_mask: true
62
+ dtype: float16
63
+ ```