mllm-dev commited on
Commit
9786b10
·
verified ·
1 Parent(s): a826ce6

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,10 +1,10 @@
1
  ---
2
  base_model:
3
- - mllm-dev/gpt2_f_experiment_3
4
  - mllm-dev/gpt2_f_experiment_4
5
- - mllm-dev/gpt2_f_experiment_1
6
  - mllm-dev/gpt2_f_experiment_0
7
- - mllm-dev/gpt2_f_experiment_2
8
  library_name: transformers
9
  tags:
10
  - mergekit
@@ -23,10 +23,10 @@ This model was merged using the linear [DARE](https://arxiv.org/abs/2311.03099)
23
  ### Models Merged
24
 
25
  The following models were included in the merge:
26
- * [mllm-dev/gpt2_f_experiment_3](https://huggingface.co/mllm-dev/gpt2_f_experiment_3)
27
  * [mllm-dev/gpt2_f_experiment_4](https://huggingface.co/mllm-dev/gpt2_f_experiment_4)
 
28
  * [mllm-dev/gpt2_f_experiment_1](https://huggingface.co/mllm-dev/gpt2_f_experiment_1)
29
- * [mllm-dev/gpt2_f_experiment_2](https://huggingface.co/mllm-dev/gpt2_f_experiment_2)
30
 
31
  ### Configuration
32
 
@@ -53,23 +53,23 @@ slices:
53
  model:
54
  path: mllm-dev/gpt2_f_experiment_1
55
  parameters:
56
- weight: 1.0
57
  - layer_range: [0, 12]
58
  model:
59
  model:
60
  path: mllm-dev/gpt2_f_experiment_2
61
  parameters:
62
- weight: 1.0
63
  - layer_range: [0, 12]
64
  model:
65
  model:
66
  path: mllm-dev/gpt2_f_experiment_3
67
  parameters:
68
- weight: 1.0
69
  - layer_range: [0, 12]
70
  model:
71
  model:
72
  path: mllm-dev/gpt2_f_experiment_4
73
  parameters:
74
- weight: 1.0
75
  ```
 
1
  ---
2
  base_model:
3
+ - mllm-dev/gpt2_f_experiment_2
4
  - mllm-dev/gpt2_f_experiment_4
5
+ - mllm-dev/gpt2_f_experiment_3
6
  - mllm-dev/gpt2_f_experiment_0
7
+ - mllm-dev/gpt2_f_experiment_1
8
  library_name: transformers
9
  tags:
10
  - mergekit
 
23
  ### Models Merged
24
 
25
  The following models were included in the merge:
26
+ * [mllm-dev/gpt2_f_experiment_2](https://huggingface.co/mllm-dev/gpt2_f_experiment_2)
27
  * [mllm-dev/gpt2_f_experiment_4](https://huggingface.co/mllm-dev/gpt2_f_experiment_4)
28
+ * [mllm-dev/gpt2_f_experiment_3](https://huggingface.co/mllm-dev/gpt2_f_experiment_3)
29
  * [mllm-dev/gpt2_f_experiment_1](https://huggingface.co/mllm-dev/gpt2_f_experiment_1)
 
30
 
31
  ### Configuration
32
 
 
53
  model:
54
  path: mllm-dev/gpt2_f_experiment_1
55
  parameters:
56
+ weight: 1.5
57
  - layer_range: [0, 12]
58
  model:
59
  model:
60
  path: mllm-dev/gpt2_f_experiment_2
61
  parameters:
62
+ weight: 1.5
63
  - layer_range: [0, 12]
64
  model:
65
  model:
66
  path: mllm-dev/gpt2_f_experiment_3
67
  parameters:
68
+ weight: 0.5
69
  - layer_range: [0, 12]
70
  model:
71
  model:
72
  path: mllm-dev/gpt2_f_experiment_4
73
  parameters:
74
+ weight: 0.5
75
  ```
mergekit_config.yml CHANGED
@@ -18,22 +18,22 @@ slices:
18
  model:
19
  path: mllm-dev/gpt2_f_experiment_1
20
  parameters:
21
- weight: 1.0
22
  - layer_range: [0, 12]
23
  model:
24
  model:
25
  path: mllm-dev/gpt2_f_experiment_2
26
  parameters:
27
- weight: 1.0
28
  - layer_range: [0, 12]
29
  model:
30
  model:
31
  path: mllm-dev/gpt2_f_experiment_3
32
  parameters:
33
- weight: 1.0
34
  - layer_range: [0, 12]
35
  model:
36
  model:
37
  path: mllm-dev/gpt2_f_experiment_4
38
  parameters:
39
- weight: 1.0
 
18
  model:
19
  path: mllm-dev/gpt2_f_experiment_1
20
  parameters:
21
+ weight: 1.5
22
  - layer_range: [0, 12]
23
  model:
24
  model:
25
  path: mllm-dev/gpt2_f_experiment_2
26
  parameters:
27
+ weight: 1.5
28
  - layer_range: [0, 12]
29
  model:
30
  model:
31
  path: mllm-dev/gpt2_f_experiment_3
32
  parameters:
33
+ weight: 0.5
34
  - layer_range: [0, 12]
35
  model:
36
  model:
37
  path: mllm-dev/gpt2_f_experiment_4
38
  parameters:
39
+ weight: 0.5
model-00001-of-00001.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab0eaa17afae2bcadd6592c25f8adadb713e147a7f00c7f0adebb89d6627b093
3
  size 248902264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96a38b8a12ea7399d70dba59681cb7edf01527d7719518377a97e2eb519964be
3
  size 248902264