bartowski commited on
Commit
b8cebb4
·
1 Parent(s): c0d9123

Main branch

Browse files
Files changed (1) hide show
  1. README.md +42 -0
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ language:
4
+ - en
5
+ pipeline_tag: text-generation
6
+ quantized_by: bartowski
7
+ ---
8
+
9
+ ## Exllama v2 Quantizations of NeuralPivot-Mistral-13B-experimental
10
+
11
+ Using <a href="https://github.com/turboderp/exllamav2/tree/experimental">turboderp's ExLlamaV2 experimental</a> for quantization.
12
+
13
+ Each branch contains an individual bits per weight. This is an experimental ExLlamaV2 quantization with no measurement.json produced.
14
+
15
+ Default arguments used except when the bits per weight is above 6.0, at that point the lm_head layer is quantized at 8 bits per weight instead of the default 6.
16
+
17
+ Original model: https://huggingface.co/athirdpath/NeuralPivot-Mistral-13B
18
+
19
+
20
+
21
+ <a href="https://huggingface.co/bartowski/NeuralPivot-Mistral-13B-experimental-exl2/tree/3_75">3.75 bits per weight</a>
22
+
23
+ ## Download instructions
24
+
25
+ With git:
26
+
27
+ ```shell
28
+ git clone --single-branch --branch 4_0 https://huggingface.co/bartowski/NeuralPivot-Mistral-13B-experimental-exl2
29
+ ```
30
+
31
+ With huggingface hub (credit to TheBloke for instructions):
32
+
33
+ ```shell
34
+ pip3 install huggingface-hub
35
+ ```
36
+
37
+ To download a specific BPW branch to a folder called `NeuralPivot-Mistral-13B-experimental-exl2`:
38
+
39
+ ```shell
40
+ mkdir NeuralPivot-Mistral-13B-experimental-exl2
41
+ huggingface-cli download bartowski/NeuralPivot-Mistral-13B-experimental-exl2 --revision 4_0 --local-dir NeuralPivot-Mistral-13B-experimental-exl2 --local-dir-use-symlinks False
42
+ ```