Sumail commited on
Commit
18e1519
·
verified ·
1 Parent(s): 60344c8

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  base_model:
 
3
  - GamblerOnTrain/danke20a
4
- - coffie3/s32
5
  library_name: transformers
6
  tags:
7
  - mergekit
@@ -20,8 +20,8 @@ This model was merged using the SLERP merge method.
20
  ### Models Merged
21
 
22
  The following models were included in the merge:
 
23
  * [GamblerOnTrain/danke20a](https://huggingface.co/GamblerOnTrain/danke20a)
24
- * [coffie3/s32](https://huggingface.co/coffie3/s32)
25
 
26
  ### Configuration
27
 
@@ -33,17 +33,17 @@ slices:
33
  - sources:
34
  - model: GamblerOnTrain/danke20a
35
  layer_range: [0, 24]
36
- - model: coffie3/s32
37
  layer_range: [0, 24]
38
  merge_method: slerp
39
- base_model: GamblerOnTrain/danke20a
40
  parameters:
41
  t:
42
  - filter: self_attn
43
  value: [0, 0.5, 0.3, 0.7, 1]
44
  - filter: mlp
45
  value: [1, 0.5, 0.7, 0.3, 0]
46
- - value: 0.35
47
  dtype: bfloat16
48
 
49
  ```
 
1
  ---
2
  base_model:
3
+ - GamblerOnTrain/danke30a
4
  - GamblerOnTrain/danke20a
 
5
  library_name: transformers
6
  tags:
7
  - mergekit
 
20
  ### Models Merged
21
 
22
  The following models were included in the merge:
23
+ * [GamblerOnTrain/danke30a](https://huggingface.co/GamblerOnTrain/danke30a)
24
  * [GamblerOnTrain/danke20a](https://huggingface.co/GamblerOnTrain/danke20a)
 
25
 
26
  ### Configuration
27
 
 
33
  - sources:
34
  - model: GamblerOnTrain/danke20a
35
  layer_range: [0, 24]
36
+ - model: GamblerOnTrain/danke30a
37
  layer_range: [0, 24]
38
  merge_method: slerp
39
+ base_model: GamblerOnTrain/danke30a
40
  parameters:
41
  t:
42
  - filter: self_attn
43
  value: [0, 0.5, 0.3, 0.7, 1]
44
  - filter: mlp
45
  value: [1, 0.5, 0.7, 0.3, 0]
46
+ - value: 0.5
47
  dtype: bfloat16
48
 
49
  ```
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "GamblerOnTrain/danke20a",
3
  "architectures": [
4
  "StableLmForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "GamblerOnTrain/danke30a",
3
  "architectures": [
4
  "StableLmForCausalLM"
5
  ],
mergekit_config.yml CHANGED
@@ -3,15 +3,15 @@ slices:
3
  - sources:
4
  - model: GamblerOnTrain/danke20a
5
  layer_range: [0, 24]
6
- - model: coffie3/s32
7
  layer_range: [0, 24]
8
  merge_method: slerp
9
- base_model: GamblerOnTrain/danke20a
10
  parameters:
11
  t:
12
  - filter: self_attn
13
  value: [0, 0.5, 0.3, 0.7, 1]
14
  - filter: mlp
15
  value: [1, 0.5, 0.7, 0.3, 0]
16
- - value: 0.35
17
  dtype: bfloat16
 
3
  - sources:
4
  - model: GamblerOnTrain/danke20a
5
  layer_range: [0, 24]
6
+ - model: GamblerOnTrain/danke30a
7
  layer_range: [0, 24]
8
  merge_method: slerp
9
+ base_model: GamblerOnTrain/danke30a
10
  parameters:
11
  t:
12
  - filter: self_attn
13
  value: [0, 0.5, 0.3, 0.7, 1]
14
  - filter: mlp
15
  value: [1, 0.5, 0.7, 0.3, 0]
16
+ - value: 0.5
17
  dtype: bfloat16
model-00001-of-00001.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8764e607696766aadef88d53e0f2df16a70e1a45237464d4390935e92d496bff
3
  size 3289069520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14ea8a6e63440a2159477deb5b99b2c329fd45c9cca51b9d68304177cb9581b9
3
  size 3289069520