Request for guff conversation or version for amd gpus

#1
by Swastikjatt7 - opened

Bro can you convert the model from fp8 to guff cause I am using amd gpu clusters which don't have native support for fp8 quantisation so can you convert fp8 to guff using llama cpp i tried but failed bcz of system specs

Sign up or log in to comment