grimjim commited on
Commit
24a3609
·
verified ·
1 Parent(s): 7905e33

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -42
README.md CHANGED
@@ -1,42 +1,42 @@
1
- ---
2
- base_model:
3
- - Gryphe/Codex-24B-Small-3.2
4
- - mrfakename/mistral-small-3.1-24b-base-2503-hf
5
- library_name: transformers
6
- pipeline_tag: text-generation
7
- tags:
8
- - mergekit
9
- - merge
10
- license: apache-2.0
11
- ---
12
- # demi-Codex-24B-Small-3.2b
13
-
14
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
-
16
- ## Merge Details
17
- ### Merge Method
18
-
19
- This model was merged using the [SLERP](https://en.wikipedia.org/wiki/Slerp) merge method.
20
-
21
- ### Models Merged
22
-
23
- The following models were included in the merge:
24
- * [Gryphe/Codex-24B-Small-3.2](https://huggingface.co/Gryphe/Codex-24B-Small-3.2)
25
- * [mrfakename/mistral-small-3.1-24b-base-2503-hf](https://huggingface.co/mrfakename/mistral-small-3.1-24b-base-2503-hf)
26
-
27
- ### Configuration
28
-
29
- The following YAML configuration was used to produce this model:
30
-
31
- ```yaml
32
- models:
33
- - model: mrfakename/mistral-small-3.1-24b-base-2503-hf
34
- - model: Gryphe/Codex-24B-Small-3.2
35
- merge_method: slerp
36
- base_model: mrfakename/mistral-small-3.1-24b-base-2503-hf
37
- parameters:
38
- t:
39
- - value: 0.5
40
- dtype: bfloat16
41
-
42
- ```
 
1
+ ---
2
+ base_model:
3
+ - Gryphe/Codex-24B-Small-3.2
4
+ - mrfakename/mistral-small-3.1-24b-base-2503-hf
5
+ library_name: transformers
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+ license: apache-2.0
11
+ ---
12
+ # demi-Codex-24B-Small-3.2
13
+
14
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
+
16
+ ## Merge Details
17
+ ### Merge Method
18
+
19
+ This model was merged using the [SLERP](https://en.wikipedia.org/wiki/Slerp) merge method.
20
+
21
+ ### Models Merged
22
+
23
+ The following models were included in the merge:
24
+ * [Gryphe/Codex-24B-Small-3.2](https://huggingface.co/Gryphe/Codex-24B-Small-3.2)
25
+ * [mrfakename/mistral-small-3.1-24b-base-2503-hf](https://huggingface.co/mrfakename/mistral-small-3.1-24b-base-2503-hf)
26
+
27
+ ### Configuration
28
+
29
+ The following YAML configuration was used to produce this model:
30
+
31
+ ```yaml
32
+ models:
33
+ - model: mrfakename/mistral-small-3.1-24b-base-2503-hf
34
+ - model: Gryphe/Codex-24B-Small-3.2
35
+ merge_method: slerp
36
+ base_model: mrfakename/mistral-small-3.1-24b-base-2503-hf
37
+ parameters:
38
+ t:
39
+ - value: 0.5
40
+ dtype: bfloat16
41
+
42
+ ```