Upload cuda7211/llama_cpp_python-0.3.16-cp310-cp310-linux_x86_64.whl with huggingface_hub
Browse files
.gitattributes
CHANGED
|
@@ -60,3 +60,4 @@ granite-4.0-tiny-preview-iq4_xs_T3UD.part-00002-of-00002.gguf filter=lfs diff=lf
|
|
| 60 |
llama-server-qwen3next filter=lfs diff=lfs merge=lfs -text
|
| 61 |
llama-server-7161-cuda filter=lfs diff=lfs merge=lfs -text
|
| 62 |
cuda7188/llama_cpp_python-0.3.16-cp310-cp310-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 60 |
llama-server-qwen3next filter=lfs diff=lfs merge=lfs -text
|
| 61 |
llama-server-7161-cuda filter=lfs diff=lfs merge=lfs -text
|
| 62 |
cuda7188/llama_cpp_python-0.3.16-cp310-cp310-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
| 63 |
+
cuda7211/llama_cpp_python-0.3.16-cp310-cp310-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
cuda7211/llama_cpp_python-0.3.16-cp310-cp310-linux_x86_64.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1068ef569cc5677e63e2bda842888980c014594cccf8778855f90ed2439b599c
|
| 3 |
+
size 1076270258
|