Improve language tag

#2
by lbourdois - opened
Files changed (1) hide show
  1. README.md +109 -96
README.md CHANGED
@@ -1,96 +1,109 @@
1
- ---
2
- base_model:
3
- - ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
4
- - ministral/Ministral-3b-instruct
5
- - stabilityai/stable-code-3b
6
- - PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
7
- - Bllossom/llama-3.2-Korean-Bllossom-3B
8
- - chuanli11/Llama-3.2-3B-Instruct-uncensored
9
- - Qwen/Qwen2.5-3B-Instruct
10
- - meta-llama/Llama-3.2-3B-Instruct
11
- - ibm-granite/granite-3b-code-base-2k
12
- - Qwen/Qwen2.5-3B
13
- - meta-llama/Llama-3.2-3B
14
- library_name: transformers
15
- tags:
16
- - mergekit
17
- - merge
18
-
19
- ---
20
- # merge
21
-
22
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
23
-
24
- ## Merge Details
25
- ### Merge Method
26
-
27
- This model was merged using the passthrough merge method.
28
-
29
- ### Models Merged
30
-
31
- The following models were included in the merge:
32
- * [ICEPVP8977/Uncensored_llama_3.2_3b_safetensors](https://huggingface.co/ICEPVP8977/Uncensored_llama_3.2_3b_safetensors)
33
- * [ministral/Ministral-3b-instruct](https://huggingface.co/ministral/Ministral-3b-instruct)
34
- * [stabilityai/stable-code-3b](https://huggingface.co/stabilityai/stable-code-3b)
35
- * [PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B](https://huggingface.co/PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B)
36
- * [Bllossom/llama-3.2-Korean-Bllossom-3B](https://huggingface.co/Bllossom/llama-3.2-Korean-Bllossom-3B)
37
- * [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored)
38
- * [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct)
39
- * [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
40
- * [ibm-granite/granite-3b-code-base-2k](https://huggingface.co/ibm-granite/granite-3b-code-base-2k)
41
- * [Qwen/Qwen2.5-3B](https://huggingface.co/Qwen/Qwen2.5-3B)
42
- * [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B)
43
-
44
- ### Configuration
45
-
46
- The following YAML configuration was used to produce this model:
47
-
48
- ```yaml
49
- slices:
50
- - sources:
51
- - layer_range: [0, 1]
52
- model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
53
-
54
- - sources:
55
- - layer_range: [0, 1]
56
- model: meta-llama/Llama-3.2-3B-Instruct
57
-
58
- - sources:
59
- - layer_range: [0, 1]
60
- model: meta-llama/Llama-3.2-3B
61
-
62
- - sources:
63
- - layer_range: [0, 1]
64
- model: chuanli11/Llama-3.2-3B-Instruct-uncensored
65
-
66
- - sources:
67
- - layer_range: [0, 1]
68
- model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
69
-
70
- - sources:
71
- - layer_range: [0, 1]
72
- model: Bllossom/llama-3.2-Korean-Bllossom-3B
73
-
74
- - sources:
75
- - layer_range: [0, 1]
76
- model: Qwen/Qwen2.5-3B-Instruct
77
-
78
- - sources:
79
- - layer_range: [0, 1]
80
- model: Qwen/Qwen2.5-3B
81
-
82
- - sources:
83
- - layer_range: [0, 1]
84
- model: stabilityai/stable-code-3b
85
-
86
- - sources:
87
- - layer_range: [0, 1]
88
- model: ibm-granite/granite-3b-code-base-2k
89
-
90
- - sources:
91
- - layer_range: [0, 2]
92
- model: ministral/Ministral-3b-instruct
93
-
94
- merge_method: passthrough
95
- dtype: float16
96
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
4
+ - ministral/Ministral-3b-instruct
5
+ - stabilityai/stable-code-3b
6
+ - PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
7
+ - Bllossom/llama-3.2-Korean-Bllossom-3B
8
+ - chuanli11/Llama-3.2-3B-Instruct-uncensored
9
+ - Qwen/Qwen2.5-3B-Instruct
10
+ - meta-llama/Llama-3.2-3B-Instruct
11
+ - ibm-granite/granite-3b-code-base-2k
12
+ - Qwen/Qwen2.5-3B
13
+ - meta-llama/Llama-3.2-3B
14
+ library_name: transformers
15
+ tags:
16
+ - mergekit
17
+ - merge
18
+ language:
19
+ - zho
20
+ - eng
21
+ - fra
22
+ - spa
23
+ - por
24
+ - deu
25
+ - ita
26
+ - rus
27
+ - jpn
28
+ - kor
29
+ - vie
30
+ - tha
31
+ - ara
32
+ ---
33
+ # merge
34
+
35
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
36
+
37
+ ## Merge Details
38
+ ### Merge Method
39
+
40
+ This model was merged using the passthrough merge method.
41
+
42
+ ### Models Merged
43
+
44
+ The following models were included in the merge:
45
+ * [ICEPVP8977/Uncensored_llama_3.2_3b_safetensors](https://huggingface.co/ICEPVP8977/Uncensored_llama_3.2_3b_safetensors)
46
+ * [ministral/Ministral-3b-instruct](https://huggingface.co/ministral/Ministral-3b-instruct)
47
+ * [stabilityai/stable-code-3b](https://huggingface.co/stabilityai/stable-code-3b)
48
+ * [PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B](https://huggingface.co/PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B)
49
+ * [Bllossom/llama-3.2-Korean-Bllossom-3B](https://huggingface.co/Bllossom/llama-3.2-Korean-Bllossom-3B)
50
+ * [chuanli11/Llama-3.2-3B-Instruct-uncensored](https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored)
51
+ * [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct)
52
+ * [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
53
+ * [ibm-granite/granite-3b-code-base-2k](https://huggingface.co/ibm-granite/granite-3b-code-base-2k)
54
+ * [Qwen/Qwen2.5-3B](https://huggingface.co/Qwen/Qwen2.5-3B)
55
+ * [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B)
56
+
57
+ ### Configuration
58
+
59
+ The following YAML configuration was used to produce this model:
60
+
61
+ ```yaml
62
+ slices:
63
+ - sources:
64
+ - layer_range: [0, 1]
65
+ model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
66
+
67
+ - sources:
68
+ - layer_range: [0, 1]
69
+ model: meta-llama/Llama-3.2-3B-Instruct
70
+
71
+ - sources:
72
+ - layer_range: [0, 1]
73
+ model: meta-llama/Llama-3.2-3B
74
+
75
+ - sources:
76
+ - layer_range: [0, 1]
77
+ model: chuanli11/Llama-3.2-3B-Instruct-uncensored
78
+
79
+ - sources:
80
+ - layer_range: [0, 1]
81
+ model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B
82
+
83
+ - sources:
84
+ - layer_range: [0, 1]
85
+ model: Bllossom/llama-3.2-Korean-Bllossom-3B
86
+
87
+ - sources:
88
+ - layer_range: [0, 1]
89
+ model: Qwen/Qwen2.5-3B-Instruct
90
+
91
+ - sources:
92
+ - layer_range: [0, 1]
93
+ model: Qwen/Qwen2.5-3B
94
+
95
+ - sources:
96
+ - layer_range: [0, 1]
97
+ model: stabilityai/stable-code-3b
98
+
99
+ - sources:
100
+ - layer_range: [0, 1]
101
+ model: ibm-granite/granite-3b-code-base-2k
102
+
103
+ - sources:
104
+ - layer_range: [0, 2]
105
+ model: ministral/Ministral-3b-instruct
106
+
107
+ merge_method: passthrough
108
+ dtype: float16
109
+ ```