view reply Great news. Serving with llama.cpp using HF-hosted models, including unsloth's on AMD Strix Halo and OpenCode here.
view article Article GGML and llama.cpp join HF to ensure the long-term progress of Local AI +4 2 days ago • 232