lbourdois commited on
Commit
a28e14b
·
verified ·
1 Parent(s): c5413a4

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +103 -90
README.md CHANGED
@@ -1,90 +1,103 @@
1
- ---
2
- base_model:
3
- - ibm-granite/granite-3b-code-base-2k
4
- - PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
5
- - stabilityai/stable-code-3b
6
- - Bllossom/llama-3.2-Korean-Bllossom-3B
7
- - ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
8
- - Qwen/Qwen2.5-3B
9
- - chuanli11/Llama-3.2-3B-Instruct-uncensored
10
- - Qwen/Qwen2.5-3B-Instruct
11
- - meta-llama/Llama-3.2-3B-Instruct
12
- - meta-llama/Llama-3.2-3B
13
- library_name: transformers
14
- tags:
15
- - mergekit
16
- - merge
17
-
18
- ---
19
- # merge
20
-
21
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
22
-
23
- ## Merge Details
24
- ### Merge Method
25
-
26
- This model was merged using the passthrough merge method.
27
-
28
- ### Models Merged
29
-
30
- The following models were included in the merge:
31
- * [ibm-granite/granite-3b-code-base-2k](https://huggingface.co/ibm-granite/granite-3b-code-base-2k)
32
- * [PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B](https://huggingface.co/PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B)
33
- * [stabilityai/stable-code-3b](https://huggingface.co/stabilityai/stable-code-3b)
34
- * [Bllossom/llama-3.2-Korean-Bllossom-3B](https://huggingface.co/Bllossom/llama-3.2-Korean-Bllossom-3B)
35
- * [ICEPVP8977/Uncensored_llama_3.2_3b_safetensors](https://huggingface.co/ICEPVP8977/Uncensored_llama_3.2_3b_safetensors)
36
- * [Qwen/Qwen2.5-3B](https://huggingface.co/Qwen/Qwen2.5-3B)
37
- * [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored)
38
- * [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct)
39
- * [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
40
- * [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B)
41
-
42
- ### Configuration
43
-
44
- The following YAML configuration was used to produce this model:
45
-
46
- ```yaml
47
- slices:
48
- - sources:
49
- - layer_range: [0, 1]
50
- model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
51
-
52
- - sources:
53
- - layer_range: [0, 1]
54
- model: meta-llama/Llama-3.2-3B-Instruct
55
-
56
- - sources:
57
- - layer_range: [0, 1]
58
- model: meta-llama/Llama-3.2-3B
59
-
60
- - sources:
61
- - layer_range: [0, 1]
62
- model: chuanli11/Llama-3.2-3B-Instruct-uncensored
63
-
64
- - sources:
65
- - layer_range: [0, 1]
66
- model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
67
-
68
- - sources:
69
- - layer_range: [0, 1]
70
- model: Bllossom/llama-3.2-Korean-Bllossom-3B
71
-
72
- - sources:
73
- - layer_range: [0, 1]
74
- model: Qwen/Qwen2.5-3B-Instruct
75
-
76
- - sources:
77
- - layer_range: [0, 1]
78
- model: Qwen/Qwen2.5-3B
79
-
80
- - sources:
81
- - layer_range: [0, 1]
82
- model: stabilityai/stable-code-3b
83
-
84
- - sources:
85
- - layer_range: [0, 1]
86
- model: ibm-granite/granite-3b-code-base-2k
87
-
88
- merge_method: passthrough
89
- dtype: float16
90
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - ibm-granite/granite-3b-code-base-2k
4
+ - PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
5
+ - stabilityai/stable-code-3b
6
+ - Bllossom/llama-3.2-Korean-Bllossom-3B
7
+ - ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
8
+ - Qwen/Qwen2.5-3B
9
+ - chuanli11/Llama-3.2-3B-Instruct-uncensored
10
+ - Qwen/Qwen2.5-3B-Instruct
11
+ - meta-llama/Llama-3.2-3B-Instruct
12
+ - meta-llama/Llama-3.2-3B
13
+ library_name: transformers
14
+ tags:
15
+ - mergekit
16
+ - merge
17
+ language:
18
+ - zho
19
+ - eng
20
+ - fra
21
+ - spa
22
+ - por
23
+ - deu
24
+ - ita
25
+ - rus
26
+ - jpn
27
+ - kor
28
+ - vie
29
+ - tha
30
+ - ara
31
+ ---
32
+ # merge
33
+
34
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
35
+
36
+ ## Merge Details
37
+ ### Merge Method
38
+
39
+ This model was merged using the passthrough merge method.
40
+
41
+ ### Models Merged
42
+
43
+ The following models were included in the merge:
44
+ * [ibm-granite/granite-3b-code-base-2k](https://huggingface.co/ibm-granite/granite-3b-code-base-2k)
45
+ * [PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B](https://huggingface.co/PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B)
46
+ * [stabilityai/stable-code-3b](https://huggingface.co/stabilityai/stable-code-3b)
47
+ * [Bllossom/llama-3.2-Korean-Bllossom-3B](https://huggingface.co/Bllossom/llama-3.2-Korean-Bllossom-3B)
48
+ * [ICEPVP8977/Uncensored_llama_3.2_3b_safetensors](https://huggingface.co/ICEPVP8977/Uncensored_llama_3.2_3b_safetensors)
49
+ * [Qwen/Qwen2.5-3B](https://huggingface.co/Qwen/Qwen2.5-3B)
50
+ * [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored)
51
+ * [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct)
52
+ * [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
53
+ * [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B)
54
+
55
+ ### Configuration
56
+
57
+ The following YAML configuration was used to produce this model:
58
+
59
+ ```yaml
60
+ slices:
61
+ - sources:
62
+ - layer_range: [0, 1]
63
+ model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
64
+
65
+ - sources:
66
+ - layer_range: [0, 1]
67
+ model: meta-llama/Llama-3.2-3B-Instruct
68
+
69
+ - sources:
70
+ - layer_range: [0, 1]
71
+ model: meta-llama/Llama-3.2-3B
72
+
73
+ - sources:
74
+ - layer_range: [0, 1]
75
+ model: chuanli11/Llama-3.2-3B-Instruct-uncensored
76
+
77
+ - sources:
78
+ - layer_range: [0, 1]
79
+ model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
80
+
81
+ - sources:
82
+ - layer_range: [0, 1]
83
+ model: Bllossom/llama-3.2-Korean-Bllossom-3B
84
+
85
+ - sources:
86
+ - layer_range: [0, 1]
87
+ model: Qwen/Qwen2.5-3B-Instruct
88
+
89
+ - sources:
90
+ - layer_range: [0, 1]
91
+ model: Qwen/Qwen2.5-3B
92
+
93
+ - sources:
94
+ - layer_range: [0, 1]
95
+ model: stabilityai/stable-code-3b
96
+
97
+ - sources:
98
+ - layer_range: [0, 1]
99
+ model: ibm-granite/granite-3b-code-base-2k
100
+
101
+ merge_method: passthrough
102
+ dtype: float16
103
+ ```