This is a merge of pre-trained language models created using mergekit. This model was merged using the Karcher merge method.
Goetia version 1.2 (Checkpoint S) represents a major upgrade over v1.1. Eighteen models were combined for this behemoth merge. The following changes were made to the Goetic pipeline:
- No merges were used as donors. Finetunes only, as with the original Cthulhu. This offers the least amount of vector distortion and highest accuracy for the PCA manifold. The graph_v18.py script helped tremendously to merge with a 3060 Ti.
- All 2501 finetunes were removed due to incompatibility. Only MS 2503/2506 finetunes were added. Boreas is basically the 'outtakes' version of Goetia, featuring Mistral 2501 finetunes.
- Custom methods like
fluxandchiral_qhehave been developed but are still being refined.karcherwas chosen because it produces the most stable merge using normal methods for 10+ donors. - Goetia was originally intended to be the spiritual successor to the Cthulhu series, made using mergekit. Now however, it's considered to be a checkpoint for Cthulhu v1.4, which is planned to be a finetune of the latest Goetia 24B. The goal is to uncensor it during finetuning so that ablation isn't needed, and to train it on H.P. Lovecraft datasets. See Avnas 7B for a preview of this.
- Check out Soulblighter or StormSeeker for Goetia alternatives.
- team mradermacher and seeingterra for GGUF
- DeathGodlike and ArtusDev for EXL3
- McG-221 for MLX
Due to storage limits only select GGUFs are available. IQ quants were made with illuminati_imatrix_v1.txt.
Download GGUFs: Naphula/Goetia-24B-v1.2-GGUF
| Quant | Size |
|---|---|
| IQ1_S | 5.27 GB |
| IQ2_M | 8.11 GB |
| IQ3_M | 10.7 GB |
| IQ4_XS | 12.8 GB |
| IQ4_NL | 13.5 GB |
| Q5_K_M | 16.8 GB |
| Q6_K | 19.3 GB |
| Q8_K_XL | 29.8 GB |
I recommend these pages for other quantizations:
>

