Improve language tag

#2
by lbourdois - opened
Files changed (1) hide show
  1. README.md +64 -64
README.md CHANGED
@@ -1,64 +1,64 @@
1
- ---
2
- base_model:
3
- - Qwen/Qwen2.5-14B
4
- - Qwen/Qwen2.5-14B-Instruct
5
- - qingy2019/Qwen2.5-Math-14B-Instruct-Alpha
6
- library_name: transformers
7
- tags:
8
- - mergekit
9
- - merge
10
- language:
11
- - zho
12
- - eng
13
- - fra
14
- - spa
15
- - por
16
- - deu
17
- - ita
18
- - rus
19
- - jpn
20
- - kor
21
- - vie
22
- - tha
23
- - ara
24
- ---
25
- # merge
26
-
27
- This is a merge of pre-trained language models created using mergekit
28
-
29
- ## Merge Details
30
- ### Merge Method
31
-
32
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [Qwen/Qwen2.5-14B](https://huggingface.co/Qwen/Qwen2.5-14B) as a base.
33
-
34
- ### Models Merged
35
-
36
- The following models were included in the merge:
37
- * [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct)
38
- * [qingy2019/Qwen2.5-Math-14B-Instruct-Alpha](https://huggingface.co/qingy2019/Qwen2.5-Math-14B-Instruct-Alpha)
39
-
40
- ### Configuration
41
-
42
- The following YAML configuration was used to produce this model:
43
-
44
- ```yaml
45
- models:
46
- - model: qingy2019/Qwen2.5-Math-14B-Instruct-Alpha
47
- parameters:
48
- weight: 1
49
- density: 1
50
- - model: Qwen/Qwen2.5-14B-Instruct
51
- parameters:
52
- weight: 1
53
- density: 1
54
- merge_method: ties
55
- base_model: Qwen/Qwen2.5-14B
56
- parameters:
57
- weight: 1
58
- density: 1
59
- normalize: true
60
- int8_mask: true
61
- tokenizer_source: qingy2019/Qwen2.5-Math-14B-Instruct-Alpha
62
- dtype: bfloat16
63
-
64
- ```
 
1
+ ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-14B
4
+ - Qwen/Qwen2.5-14B-Instruct
5
+ - qingy2019/Qwen2.5-Math-14B-Instruct-Alpha
6
+ library_name: transformers
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+ language:
11
+ - zho
12
+ - eng
13
+ - fra
14
+ - spa
15
+ - por
16
+ - deu
17
+ - ita
18
+ - rus
19
+ - jpn
20
+ - kor
21
+ - vie
22
+ - tha
23
+ - ara
24
+ ---
25
+ # merge
26
+
27
+ This is a merge of pre-trained language models created using mergekit
28
+
29
+ ## Merge Details
30
+ ### Merge Method
31
+
32
+ This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [Qwen/Qwen2.5-14B](https://huggingface.co/Qwen/Qwen2.5-14B) as a base.
33
+
34
+ ### Models Merged
35
+
36
+ The following models were included in the merge:
37
+ * [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct)
38
+ * [qingy2019/Qwen2.5-Math-14B-Instruct-Alpha](https://huggingface.co/qingy2019/Qwen2.5-Math-14B-Instruct-Alpha)
39
+
40
+ ### Configuration
41
+
42
+ The following YAML configuration was used to produce this model:
43
+
44
+ ```yaml
45
+ models:
46
+ - model: qingy2019/Qwen2.5-Math-14B-Instruct-Alpha
47
+ parameters:
48
+ weight: 1
49
+ density: 1
50
+ - model: Qwen/Qwen2.5-14B-Instruct
51
+ parameters:
52
+ weight: 1
53
+ density: 1
54
+ merge_method: ties
55
+ base_model: Qwen/Qwen2.5-14B
56
+ parameters:
57
+ weight: 1
58
+ density: 1
59
+ normalize: true
60
+ int8_mask: true
61
+ tokenizer_source: qingy2019/Qwen2.5-Math-14B-Instruct-Alpha
62
+ dtype: bfloat16
63
+
64
+ ```