ubergarm commited on
Commit
28a18a8
·
0 Parent(s):

initial commit

Browse files
Files changed (2) hide show
  1. .gitattributes +38 -0
  2. README.md +87 -0
.gitattributes ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ imatrix-*.dat filter=lfs diff=lfs merge=lfs -text
37
+ *.gguf filter=lfs diff=lfs merge=lfs -text
38
+ *.png filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ quantized_by: ubergarm
3
+ pipeline_tag: text-generation
4
+ base_model: stepfun-ai/Step-3.5-Flash
5
+ base_model_relation: quantized
6
+ license: apache-2.0
7
+ tags:
8
+ - imatrix
9
+ - conversational
10
+ - ik_llama.cpp
11
+ - step3p5
12
+ ---
13
+
14
+ ## WIP
15
+ - [ ] compute imatrix against bf16 and upload
16
+ - [ ] cook some quants
17
+ - [ ] test perplexity
18
+
19
+ ## `ik_llama.cpp` imatrix Quantizations of stepfun-ai/Step-3.5-Flash
20
+ *NOTE* `ik_llama.cpp` can also run your existing GGUFs from bartowski, unsloth, mradermacher, etc if you want to try it out before downloading my quants.
21
+
22
+ Some of ik's new quants are supported with [Nexesenex/croco.cpp](https://github.com/Nexesenex/croco.cpp) fork of KoboldCPP with Windows builds for CUDA 12.9. Also check for [Windows builds by Thireus here.](https://github.com/Thireus/ik_llama.cpp/releases) which have been CUDA 12.8.
23
+
24
+ These quants provide best in class perplexity for the given memory footprint.
25
+
26
+ ## Big Thanks
27
+ Shout out to Wendell and the **Level1Techs** crew, the community [Forums](https://forum.level1techs.com/t/deepseek-deep-dive-r1-at-home/225826), [YouTube Channel](https://www.youtube.com/@Level1Techs)! **BIG thanks** for providing **BIG hardware** expertise and access to run these experiments and make these great quants available to the community!!!
28
+
29
+ Also thanks to all the folks in the quanting and inferencing community on [BeaverAI Club Discord](https://huggingface.co/BeaverAI) and on [r/LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/) for tips and tricks helping each other run, test, and benchmark all the fun new models! Thanks to huggingface for hosting all these big quants!
30
+
31
+ Finally, I *really* appreciate the support from [aifoundry.org](https://aifoundry.org) so check out their open source RISC-V based solutions!
32
+
33
+ ## Quant Collection
34
+ Perplexity computed against *wiki.test.raw*. (lower is "better")
35
+
36
+ ![Perplexity Chart](images/perplexity.png "Chart showing Perplexity vs Model Size.")
37
+ (not sure why the y-axis doesn't look log scale on this one)
38
+
39
+ These two are just a test quants for baseline perplexity comparison:
40
+ * `BF16` TODO
41
+ - TODO
42
+ * `Q8_0` TODO
43
+ - TODO
44
+
45
+ *NOTE*: The first split file is much smaller on purpose to only contain metadata, its fine!
46
+
47
+ ## IQ5_K TODO
48
+ TODO
49
+
50
+ <details>
51
+
52
+ <summary>👈 Secret Recipe</summary>
53
+
54
+ ```bash
55
+ echo TODO
56
+ ```
57
+
58
+ </details>
59
+
60
+ ## More Quants TODO
61
+
62
+ ## Quick Start
63
+
64
+ ```bash
65
+ # Clone and checkout
66
+ $ git clone https://github.com/ikawrakow/ik_llama.cpp
67
+ $ cd ik_llama.cpp
68
+
69
+ # Build for hybrid CPU+CUDA
70
+ $ cmake -B build -DCMAKE_BUILD_TYPE=Release -DGGML_CUDA=ON
71
+ $ cmake --build build --config Release -j $(nproc)
72
+
73
+ ## https://github.com/ikawrakow/ik_llama.cpp/pull/1236
74
+ ## https://github.com/ikawrakow/ik_llama.cpp/pull/1231
75
+ ## https://github.com/ikawrakow/ik_llama.cpp/pull/1239
76
+ ## https://github.com/ikawrakow/ik_llama.cpp/pull/1240
77
+ echo TODO
78
+ ```
79
+
80
+ For tool use you can always bring your own template with `--jinja --chat-template-file myTemplate.jinja` and might need `--special` etc.
81
+
82
+ ## References
83
+ * [ik_llama.cpp](https://github.com/ikawrakow/ik_llama.cpp)
84
+ * [Getting Started Guide (already out of date lol)](https://github.com/ikawrakow/ik_llama.cpp/discussions/258)
85
+ * [ubergarm-imatrix-calibration-corpus-v02.txt](https://gist.github.com/ubergarm/edfeb3ff9c6ec8b49e88cdf627b0711a?permalink_comment_id=5682584#gistcomment-5682584)
86
+ * [mainline llama.cpp PR19283](https://github.com/ggml-org/llama.cpp/pull/19283) converted with `pull/19283/head:pr/step3.5-flash@402fc2e4e` plus casting `step35.attention.sliding_window_pattern` to `[INT32]` as for some reason it defaults to `[BOOL]` for me.
87
+ * [ik_llama.cpp PR1231](https://github.com/ikawrakow/ik_llama.cpp/pull/1231) imatrix & quantized with `ik/step35_compat@9a0b5e80`