Upload GGUF model
Browse files- .gitattributes +1 -0
- README.md +2 -2
- keip-assistant.q8_0.gguf +3 -0
.gitattributes
CHANGED
|
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
keip-assistant.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
keip-assistant.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
keip-assistant.q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
|
@@ -6,7 +6,7 @@ This is a GGUF version of the lora-merged model.
|
|
| 6 |
|
| 7 |
- **Base Model:** /workspace/lora-merged
|
| 8 |
- **Format:** GGUF
|
| 9 |
-
- **Quantization:**
|
| 10 |
|
| 11 |
## Usage
|
| 12 |
|
|
@@ -14,5 +14,5 @@ This model can be used with [llama.cpp](https://github.com/ggerganov/llama.cpp)
|
|
| 14 |
|
| 15 |
```bash
|
| 16 |
# Example llama.cpp command
|
| 17 |
-
./main -m keip-assistant.
|
| 18 |
```
|
|
|
|
| 6 |
|
| 7 |
- **Base Model:** /workspace/lora-merged
|
| 8 |
- **Format:** GGUF
|
| 9 |
+
- **Quantization:** q8_0
|
| 10 |
|
| 11 |
## Usage
|
| 12 |
|
|
|
|
| 14 |
|
| 15 |
```bash
|
| 16 |
# Example llama.cpp command
|
| 17 |
+
./main -m keip-assistant.q8_0.gguf -n 1024 -p "Your prompt here"
|
| 18 |
```
|
keip-assistant.q8_0.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:21ac1b4f17a7f386bdbb14680262ea79ba1fa30f40459af7c51b1035c42211e0
|
| 3 |
+
size 8709518048
|