lbourdois commited on
Commit
8e2aaae
·
verified ·
1 Parent(s): f64301d

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +74 -61
README.md CHANGED
@@ -1,61 +1,74 @@
1
- ---
2
- base_model:
3
- - Qwen/Qwen2.5-7B-Instruct
4
- - Bui1dMySea/LongRAG-Qwen2.5-7B-Instruct
5
- - prithivMLmods/QwQ-MathOct-7B
6
- - Krystalan/DRT-o1-7B
7
- - prithivMLmods/QwQ-LCoT-7B-Instruct
8
- - Orion-zhen/Qwen2.5-7B-Instruct-Uncensored
9
- - Spestly/Athena-1-7B
10
- - prithivMLmods/Deepthink-Reasoning-7B
11
- - fblgit/cybertron-v4-qw7B-MGS
12
- - Rombo-Org/Rombo-LLM-V2.5-Qwen-7b
13
- library_name: transformers
14
- tags:
15
- - mergekit
16
- - merge
17
-
18
- ---
19
- # merge
20
-
21
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
22
-
23
- ## Merge Details
24
- ### Merge Method
25
-
26
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) as a base.
27
-
28
- ### Models Merged
29
-
30
- The following models were included in the merge:
31
- * [Bui1dMySea/LongRAG-Qwen2.5-7B-Instruct](https://huggingface.co/Bui1dMySea/LongRAG-Qwen2.5-7B-Instruct)
32
- * [prithivMLmods/QwQ-MathOct-7B](https://huggingface.co/prithivMLmods/QwQ-MathOct-7B)
33
- * [Krystalan/DRT-o1-7B](https://huggingface.co/Krystalan/DRT-o1-7B)
34
- * [prithivMLmods/QwQ-LCoT-7B-Instruct](https://huggingface.co/prithivMLmods/QwQ-LCoT-7B-Instruct)
35
- * [Orion-zhen/Qwen2.5-7B-Instruct-Uncensored](https://huggingface.co/Orion-zhen/Qwen2.5-7B-Instruct-Uncensored)
36
- * [Spestly/Athena-1-7B](https://huggingface.co/Spestly/Athena-1-7B)
37
- * [prithivMLmods/Deepthink-Reasoning-7B](https://huggingface.co/prithivMLmods/Deepthink-Reasoning-7B)
38
- * [fblgit/cybertron-v4-qw7B-MGS](https://huggingface.co/fblgit/cybertron-v4-qw7B-MGS)
39
- * [Rombo-Org/Rombo-LLM-V2.5-Qwen-7b](https://huggingface.co/Rombo-Org/Rombo-LLM-V2.5-Qwen-7b)
40
-
41
- ### Configuration
42
-
43
- The following YAML configuration was used to produce this model:
44
-
45
- ```yaml
46
- models:
47
- - model: Rombo-Org/Rombo-LLM-V2.5-Qwen-7b
48
- - model: fblgit/cybertron-v4-qw7B-MGS
49
- - model: Spestly/Athena-1-7B
50
- - model: Orion-zhen/Qwen2.5-7B-Instruct-Uncensored
51
- - model: prithivMLmods/Deepthink-Reasoning-7B
52
- - model: prithivMLmods/QwQ-LCoT-7B-Instruct
53
- - model: prithivMLmods/QwQ-MathOct-7B
54
- - model: Bui1dMySea/LongRAG-Qwen2.5-7B-Instruct
55
- - model: Krystalan/DRT-o1-7B
56
- merge_method: model_stock
57
- base_model: Qwen/Qwen2.5-7B-Instruct
58
- normalize: false
59
- int8_mask: true
60
- dtype: bfloat16
61
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-7B-Instruct
4
+ - Bui1dMySea/LongRAG-Qwen2.5-7B-Instruct
5
+ - prithivMLmods/QwQ-MathOct-7B
6
+ - Krystalan/DRT-o1-7B
7
+ - prithivMLmods/QwQ-LCoT-7B-Instruct
8
+ - Orion-zhen/Qwen2.5-7B-Instruct-Uncensored
9
+ - Spestly/Athena-1-7B
10
+ - prithivMLmods/Deepthink-Reasoning-7B
11
+ - fblgit/cybertron-v4-qw7B-MGS
12
+ - Rombo-Org/Rombo-LLM-V2.5-Qwen-7b
13
+ library_name: transformers
14
+ tags:
15
+ - mergekit
16
+ - merge
17
+ language:
18
+ - zho
19
+ - eng
20
+ - fra
21
+ - spa
22
+ - por
23
+ - deu
24
+ - ita
25
+ - rus
26
+ - jpn
27
+ - kor
28
+ - vie
29
+ - tha
30
+ - ara
31
+ ---
32
+ # merge
33
+
34
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
35
+
36
+ ## Merge Details
37
+ ### Merge Method
38
+
39
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) as a base.
40
+
41
+ ### Models Merged
42
+
43
+ The following models were included in the merge:
44
+ * [Bui1dMySea/LongRAG-Qwen2.5-7B-Instruct](https://huggingface.co/Bui1dMySea/LongRAG-Qwen2.5-7B-Instruct)
45
+ * [prithivMLmods/QwQ-MathOct-7B](https://huggingface.co/prithivMLmods/QwQ-MathOct-7B)
46
+ * [Krystalan/DRT-o1-7B](https://huggingface.co/Krystalan/DRT-o1-7B)
47
+ * [prithivMLmods/QwQ-LCoT-7B-Instruct](https://huggingface.co/prithivMLmods/QwQ-LCoT-7B-Instruct)
48
+ * [Orion-zhen/Qwen2.5-7B-Instruct-Uncensored](https://huggingface.co/Orion-zhen/Qwen2.5-7B-Instruct-Uncensored)
49
+ * [Spestly/Athena-1-7B](https://huggingface.co/Spestly/Athena-1-7B)
50
+ * [prithivMLmods/Deepthink-Reasoning-7B](https://huggingface.co/prithivMLmods/Deepthink-Reasoning-7B)
51
+ * [fblgit/cybertron-v4-qw7B-MGS](https://huggingface.co/fblgit/cybertron-v4-qw7B-MGS)
52
+ * [Rombo-Org/Rombo-LLM-V2.5-Qwen-7b](https://huggingface.co/Rombo-Org/Rombo-LLM-V2.5-Qwen-7b)
53
+
54
+ ### Configuration
55
+
56
+ The following YAML configuration was used to produce this model:
57
+
58
+ ```yaml
59
+ models:
60
+ - model: Rombo-Org/Rombo-LLM-V2.5-Qwen-7b
61
+ - model: fblgit/cybertron-v4-qw7B-MGS
62
+ - model: Spestly/Athena-1-7B
63
+ - model: Orion-zhen/Qwen2.5-7B-Instruct-Uncensored
64
+ - model: prithivMLmods/Deepthink-Reasoning-7B
65
+ - model: prithivMLmods/QwQ-LCoT-7B-Instruct
66
+ - model: prithivMLmods/QwQ-MathOct-7B
67
+ - model: Bui1dMySea/LongRAG-Qwen2.5-7B-Instruct
68
+ - model: Krystalan/DRT-o1-7B
69
+ merge_method: model_stock
70
+ base_model: Qwen/Qwen2.5-7B-Instruct
71
+ normalize: false
72
+ int8_mask: true
73
+ dtype: bfloat16
74
+ ```