--- base_model: - TareksLab/Carnelian-SCE-V4-LLaMa-70B - TareksLab/Emerald-SCE-V3-LLaMa-70B - TareksLab/Amethyst-SCE-V4-LLaMa-70B - TareksLab/Citrine-MS-V3-LLaMa-70B - TareksLab/Diamond-DL-V1-LLaMa-70B - TareksLab/Ruby-D-V3-LLaMa-70B library_name: transformers tags: - mergekit - merge license: llama3.3 --- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64909c086073a0cd172d0411/X9pL9M9T_oZ46n3c6aejO.png) Scripturient is a culmination of my ongoing experiments with merging specialized curated models. Designed to keep creativity high, without sacrificing stability. As for samplers, the model doesn't need samplers to reign it in much at all. My recommendation is: ``` Temp: 1 Min P: 0.01 ``` That being said, it can handle even higher temperatures and Nsigma works well too. Because of the nature of this sort of 'Hyper Multi Model Merge', my recommendation is not to run this on anything lower than a Q5 quant. If you enjoy my work, please consider supporting me, It helps me make more models like this! Support on KO-FI <3 I want to say a special thank you to everyone at the BeaverAI community who supports me, be that with testing, feedback, advice or donations! Special shoutouts to (forgive me if I left someone out!): @Artus | @Geechan | @Kromeurus | @NarpasSword | @Thana Alt | @FrenzyBiscuit | @Saintonan | @Lightning_missile | @Inasity | @Amp | @madison 🦋 @ IQ3_XS | @zerofata ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: TareksLab/Diamond-DL-V1-LLaMa-70B parameters: weight: 0.10 density: 0.7 epsilon: 0.20 - model: TareksLab/Citrine-MS-V3-LLaMa-70B parameters: weight: [0.5, 0.2, 0.1, 0.1, 0.1] density: 0.7 epsilon: 0.20 - model: TareksLab/Amethyst-SCE-V4-LLaMa-70B parameters: weight: [0.2, 0.4, 0.2, 0.1, 0.1] density: 0.7 epsilon: 0.20 - model: TareksLab/Ruby-D-V3-LLaMa-70B parameters: weight: [0.1, 0.2, 0.4, 0.2, 0.1] density: 0.7 epsilon: 0.20 - model: TareksLab/Carnelian-SCE-V4-LLaMa-70B parameters: weight: [0.1, 0.1, 0.2, 0.4, 0.2] density: 0.7 epsilon: 0.20 - model: TareksLab/Emerald-SCE-V3-LLaMa-70B parameters: weight: [0.1, 0.1, 0.1, 0.2, 0.5] density: 0.7 epsilon: 0.20 merge_method: della_linear base_model: TareksLab/Diamond-DL-V1-LLaMa-70B parameters: lambda: 1.1 normalize: false dtype: float32 out_dtype: bfloat16 chat_template: llama3 tokenizer: source: TareksLab/Ruby-D-V3-LLaMa-70B pad_to_multiple_of: 8 ```