Upload cuda7188/llama_cpp_python-0.3.16-cp310-cp310-linux_x86_64.whl with huggingface_hub
Browse files
.gitattributes
CHANGED
|
@@ -59,3 +59,4 @@ Tiny-Moe.Q6_K_T3.part-00002-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
|
|
| 59 |
granite-4.0-tiny-preview-iq4_xs_T3UD.part-00002-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
|
| 60 |
llama-server-qwen3next filter=lfs diff=lfs merge=lfs -text
|
| 61 |
llama-server-7161-cuda filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 59 |
granite-4.0-tiny-preview-iq4_xs_T3UD.part-00002-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
|
| 60 |
llama-server-qwen3next filter=lfs diff=lfs merge=lfs -text
|
| 61 |
llama-server-7161-cuda filter=lfs diff=lfs merge=lfs -text
|
| 62 |
+
cuda7188/llama_cpp_python-0.3.16-cp310-cp310-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
cuda7188/llama_cpp_python-0.3.16-cp310-cp310-linux_x86_64.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3284908ed33d6514d072865fdf33e3e15ac3b71e1a8d34fef6ea9e0b016822e2
|
| 3 |
+
size 1076124958
|