Upload with huggingface_hub
Browse files- .gitattributes +4 -0
- README.md +34 -0
- ggml-model-q4_0.bin.4 +3 -0
- ggml-model-q4_0.bin.5 +3 -0
- ggml-model-q4_0.bin.6 +3 -0
- ggml-model-q4_0.bin.7 +3 -0
.gitattributes
CHANGED
|
@@ -35,3 +35,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 35 |
ggml-model-q4_0.bin.1 filter=lfs diff=lfs merge=lfs -text
|
| 36 |
ggml-model-q4_0.bin.2 filter=lfs diff=lfs merge=lfs -text
|
| 37 |
ggml-model-q4_0.bin.3 filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
ggml-model-q4_0.bin.1 filter=lfs diff=lfs merge=lfs -text
|
| 36 |
ggml-model-q4_0.bin.2 filter=lfs diff=lfs merge=lfs -text
|
| 37 |
ggml-model-q4_0.bin.3 filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
ggml-model-q4_0.bin.4 filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
ggml-model-q4_0.bin.5 filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
ggml-model-q4_0.bin.6 filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
ggml-model-q4_0.bin.7 filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
|
| 2 |
+
### alpaca.cpp 65B ggml model weight
|
| 3 |
+
### make 65B ggml story
|
| 4 |
+
#### 1. clone 65B model data
|
| 5 |
+
```shell
|
| 6 |
+
git clone https://huggingface.co/datasets/nyanko7/LLaMA-65B/
|
| 7 |
+
```
|
| 8 |
+
#### 2. clone alpaca.cpp
|
| 9 |
+
```shell
|
| 10 |
+
git clone https://github.com/antimatter15/alpaca.cpp
|
| 11 |
+
```
|
| 12 |
+
#### 3. weight quantize.sh
|
| 13 |
+
```shell
|
| 14 |
+
|
| 15 |
+
mv LLaMA-65B/tokenizer.model ./
|
| 16 |
+
python convert-pth-to-ggml.py ../LLaMA-65B/ 1
|
| 17 |
+
|
| 18 |
+
cd alpaca.cpp
|
| 19 |
+
mkdir -p models/65B
|
| 20 |
+
mv ../LLaMA-65B/ggml-model-f16.bin models/65B/
|
| 21 |
+
mv ../LLaMA-65B/ggml-model-f16.bin.* models/65B/
|
| 22 |
+
bash quantize.sh 65B
|
| 23 |
+
```
|
| 24 |
+
|
| 25 |
+
#### 4. upload weight file
|
| 26 |
+
##### Upload is slower. The upload is taking almost 2 days, I decided to curve the upload
|
| 27 |
+
##### I using https://tmp.link/ as temp store
|
| 28 |
+
##### I using colab and huggingface api upload
|
| 29 |
+
|
| 30 |
+
### run
|
| 31 |
+
```shell
|
| 32 |
+
git clone https://github.com/antimatter15/
|
| 33 |
+
./chat -m alpaca.cpp_65b_ggml/ggml-model-q4_0.bin
|
| 34 |
+
```
|
ggml-model-q4_0.bin.4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5ee474051e418c5732b7949190b084d9d679db447f83c1de0d2a82daaa1a0cfa
|
| 3 |
+
size 5105954665
|
ggml-model-q4_0.bin.5
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a45aa05e7212bd6782790722d68056c5419667ea6b564ccc94bbcb8111d79b8b
|
| 3 |
+
size 5105954665
|
ggml-model-q4_0.bin.6
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a58fda714b759c28ad5e4c1d8bf8fda7b158fd5e4c4a49f851f36342fa97a105
|
| 3 |
+
size 5105954665
|
ggml-model-q4_0.bin.7
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a3540cfcbcda33c223c6b0d606034adbd78f17e0e5de1582b78795e78754f7a8
|
| 3 |
+
size 5105954665
|