James040 commited on
Commit
ff952d6
·
verified ·
1 Parent(s): 14da0d0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -11
README.md CHANGED
@@ -8,18 +8,35 @@ tags:
8
  - python-3.13
9
  ---
10
 
11
- # Llama-CPP-Python Pre-built Wheels (CPU Only)
12
 
13
- This repository provides pre-compiled Python wheels for `llama-cpp-python`, specifically optimized for **Hugging Face Spaces (Free CPU Tier)**.
14
 
15
- ## 🚀 Key Features
16
- - **Zero Compilation:** Skips the 15-minute C++ build process on HF Spaces.
17
- - **Python 3.13 Support:** Built for the latest Python version.
18
- - **Generic CPU:** Compiled with `GGML_NATIVE=OFF` to ensure compatibility with older cloud processors (no "Illegal Instruction" errors).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
- ## 🛠 Usage in HF Spaces
21
 
22
- ### Dockerfile (Recommended)
23
- Add this to your Dockerfile to install the wheel instantly:
24
- ```dockerfile
25
- RUN pip install [https://huggingface.co/Jameson040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl](https://huggingface.co/Jameson040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl)
 
8
  - python-3.13
9
  ---
10
 
11
+ # 🦙 Llama-CPP-Python Pre-built Wheels (Python 3.13)
12
 
13
+ ### The solution for Hugging Face "Build Timeout" errors on the Free CPU Tier.
14
 
15
+ If you are using **Python 3.13** on a Hugging Face Free Space, compiling `llama-cpp-python` from source usually crashes or times out. This repository provides pre-compiled **manylinux** wheels that install in seconds.
16
+
17
+ ---
18
+
19
+ ## 🚀 Why use these wheels?
20
+
21
+ * **No Compilation:** Skips the 15+ minute build process.
22
+ * **Python 3.13 Support:** Specifically built for the latest Python version.
23
+ * **Generic CPU Optimization:** Compiled with `GGML_NATIVE=OFF`. This ensures the model runs on HF's shared CPUs without "Illegal Instruction" or "Core Dump" errors.
24
+ * **Lightweight:** Only ~4.3 MB compared to the massive overhead of building from source.
25
+
26
+ ---
27
+
28
+ ## 🛠️ How to use in your HF Space
29
+
30
+ ### Option A: Using `requirements.txt`
31
+ Simply paste this direct link into your `requirements.txt` file:
32
+
33
+ ```text
34
+ [https://huggingface.co/James040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl](https://huggingface.co/James040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl)
35
+
36
+
37
+ Option B: Using a Dockerfile
38
+ If you are using a custom Docker setup, add this line:
39
+ RUN pip install [https://huggingface.co/James040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl](https://huggingface.co/James040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl)
40
 
 
41
 
42
+ 📦 Build SpecificationsThese wheels were built using a high-performance automated pipeline on GitHub.SpecificationValuePython Version3.13PlatformLinux x86_64 (Manylinux)Build FlagsGGML_NATIVE=OFF, GGML_BLAS=OFFBuild SourceJameson040/my_lama-wheels