SubtleOne commited on
Commit
c24f8a1
·
verified ·
1 Parent(s): 15ad578

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -57
README.md CHANGED
@@ -1,57 +1,57 @@
1
- ---
2
- base_model:
3
- - nbeerbower/Qwen2.5-Gutenberg-Doppel-32B
4
- - rombodawg/Rombos-LLM-V2.5-Qwen-32b
5
- - ArliAI/Qwen2.5-32B-ArliAI-RPMax-v1.3
6
- - EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2
7
- library_name: transformers
8
- tags:
9
- - mergekit
10
- - merge
11
-
12
- ---
13
- # merge
14
-
15
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
-
17
- ## Merge Details
18
- ### Merge Method
19
-
20
- This model was merged using the [DELLA](https://arxiv.org/abs/2406.11617) merge method using [rombodawg/Rombos-LLM-V2.5-Qwen-32b](https://huggingface.co/rombodawg/Rombos-LLM-V2.5-Qwen-32b) as a base.
21
-
22
- ### Models Merged
23
-
24
- The following models were included in the merge:
25
- * [nbeerbower/Qwen2.5-Gutenberg-Doppel-32B](https://huggingface.co/nbeerbower/Qwen2.5-Gutenberg-Doppel-32B)
26
- * [ArliAI/Qwen2.5-32B-ArliAI-RPMax-v1.3](https://huggingface.co/ArliAI/Qwen2.5-32B-ArliAI-RPMax-v1.3)
27
- * [EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2)
28
-
29
- ### Configuration
30
-
31
- The following YAML configuration was used to produce this model:
32
-
33
- ```yaml
34
- base_model: rombodawg/Rombos-LLM-V2.5-Qwen-32b
35
- parameters:
36
- int8_mask: true
37
- rescale: false
38
- normalize: true
39
- lambda: 1.04
40
- epsilon: 0.05
41
- dtype: bfloat16
42
- tokenizer_source: union
43
- merge_method: della
44
- models:
45
- - model: EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2
46
- parameters:
47
- weight: [0.40]
48
- density: [0.53]
49
- - model: nbeerbower/Qwen2.5-Gutenberg-Doppel-32B
50
- parameters:
51
- weight: [0.30]
52
- density: [0.53]
53
- - model: ArliAI/Qwen2.5-32B-ArliAI-RPMax-v1.3
54
- parameters:
55
- weight: [0.40]
56
- density: [0.53]
57
- ```
 
1
+ ---
2
+ base_model:
3
+ - nbeerbower/Qwen2.5-Gutenberg-Doppel-32B
4
+ - rombodawg/Rombos-LLM-V2.5-Qwen-32b
5
+ - ArliAI/Qwen2.5-32B-ArliAI-RPMax-v1.3
6
+ - EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2
7
+ library_name: transformers
8
+ tags:
9
+ - mergekit
10
+ - merge
11
+ license: apache-2.0
12
+ ---
13
+ # merge
14
+
15
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
+
17
+ ## Merge Details
18
+ ### Merge Method
19
+
20
+ This model was merged using the [DELLA](https://arxiv.org/abs/2406.11617) merge method using [rombodawg/Rombos-LLM-V2.5-Qwen-32b](https://huggingface.co/rombodawg/Rombos-LLM-V2.5-Qwen-32b) as a base.
21
+
22
+ ### Models Merged
23
+
24
+ The following models were included in the merge:
25
+ * [nbeerbower/Qwen2.5-Gutenberg-Doppel-32B](https://huggingface.co/nbeerbower/Qwen2.5-Gutenberg-Doppel-32B)
26
+ * [ArliAI/Qwen2.5-32B-ArliAI-RPMax-v1.3](https://huggingface.co/ArliAI/Qwen2.5-32B-ArliAI-RPMax-v1.3)
27
+ * [EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2)
28
+
29
+ ### Configuration
30
+
31
+ The following YAML configuration was used to produce this model:
32
+
33
+ ```yaml
34
+ base_model: rombodawg/Rombos-LLM-V2.5-Qwen-32b
35
+ parameters:
36
+ int8_mask: true
37
+ rescale: false
38
+ normalize: true
39
+ lambda: 1.04
40
+ epsilon: 0.05
41
+ dtype: bfloat16
42
+ tokenizer_source: union
43
+ merge_method: della
44
+ models:
45
+ - model: EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2
46
+ parameters:
47
+ weight: [0.40]
48
+ density: [0.53]
49
+ - model: nbeerbower/Qwen2.5-Gutenberg-Doppel-32B
50
+ parameters:
51
+ weight: [0.30]
52
+ density: [0.53]
53
+ - model: ArliAI/Qwen2.5-32B-ArliAI-RPMax-v1.3
54
+ parameters:
55
+ weight: [0.40]
56
+ density: [0.53]
57
+ ```