# Use CPU wheels for llama-cpp-python (avoid building from source) --prefer-binary --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cpu llama-cpp-python==0.2.90 gradio>=4.0.0 huggingface_hub>=0.20.0