Request for guff conversation or version for amd gpus
#1
by
Swastikjatt7 - opened
Bro can you convert the model from fp8 to guff cause I am using amd gpu clusters which don't have native support for fp8 quantisation so can you convert fp8 to guff using llama cpp i tried but failed bcz of system specs