view reply Great news. Serving with llama.cpp using HF-hosted models, including unsloth's on AMD Strix Halo and OpenCode here.
view article Article GGML and llama.cpp join HF to ensure the long-term progress of Local AI +4 1 day ago • 166