Naphula commited on
Commit
687f349
·
verified ·
1 Parent(s): ceea7b9

Upload mergekit_config.yml

Browse files
Files changed (1) hide show
  1. mergekit_config.yml +67 -0
mergekit_config.yml ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ architecture: MistralForCausalLM
2
+ merge_method: scream # Similarity-Consensus Resolved Enhanced Adaptive Merging
3
+ base_model: B:\24B\!models--mistralai--Magistral-Small-2509
4
+ models:
5
+ - model: B:\24B\!BeaverAI_Fallen-Mistral-Small-3.1-24B-v1e_textonly
6
+ parameters:
7
+ weight: 0.1
8
+ - model: B:\24B\!models--Casual-Autopsy--Maginum-Cydoms-24B
9
+ parameters:
10
+ weight: 0.1
11
+ - model: B:\24B\!models--DarkArtsForge--Asmodeus-24B-v1
12
+ parameters:
13
+ weight: 0.1
14
+ - model: B:\24B\!models--DarkArtsForge--Magistaroth-24B-v1
15
+ parameters:
16
+ weight: 0.1
17
+ - model: B:\24B\!models--FlareRebellion--WeirdCompound-v1.7-24b
18
+ parameters:
19
+ weight: 0.1
20
+ - model: B:\24B\!models--Gryphe--Tiamat-24B-Magistral
21
+ parameters:
22
+ weight: 0.1
23
+ - model: B:\24B\!models--Naphula--Slimaki-24B-v1
24
+ parameters:
25
+ weight: 0.1
26
+ - model: B:\24B\!models--ReadyArt--4.2.0-Broken-Tutu-24b
27
+ parameters:
28
+ weight: 0.1
29
+ - model: B:\24B\!models--sophosympatheia--Magistry-24B-v1.0
30
+ parameters:
31
+ weight: 0.1
32
+ - model: B:\24B\!models--TheDrummer--Cydonia-24B-v4.3
33
+ parameters:
34
+ weight: 0.1
35
+ - model: B:\24B\!models--TheDrummer--Magidonia-24B-v4.3
36
+ parameters:
37
+ weight: 0.1
38
+ - model: B:\24B\!models--TheDrummer--Precog-24B-v1
39
+ parameters:
40
+ weight: 0.1
41
+ - model: B:\24B\!models--zerofata--MS3.2-PaintedFantasy-v2-24B
42
+ parameters:
43
+ weight: 0.1
44
+ - model: B:\24B\!models--zerofata--MS3.2-PaintedFantasy-v3-24B
45
+ parameters:
46
+ weight: 0.1
47
+ - model: B:\24B\!models--zerofata--MS3.2-PaintedFantasy-v4.1-24B
48
+ parameters:
49
+ weight: 0.1
50
+ - model: B:\24B\!models--Naphula--Goetia-24B-v1.3
51
+ parameters:
52
+ weight: 0.1
53
+ parameters:
54
+ stock_weight: 0.4 # Weight for model_stock component
55
+ della_novelty_weight: 0.3 # Weight for DELLA novelty
56
+ sce_novelty_weight: 0.3 # Weight for SCE novelty
57
+ density: 0.9 # DELLA density parameter
58
+ epsilon: 0.05 # DELLA epsilon parameter
59
+ select_topk: 0.5 # SCE top-k selection
60
+ filter_wise: false # Model Stock filter-wise calculation
61
+ int8_mask: false # Use int8 masks for memory efficiency
62
+ dtype: float32
63
+ out_dtype: bfloat16
64
+ tokenizer:
65
+ source: union
66
+ # chat_template: auto
67
+ name: 👻 GhostFace-24B-v1