Even the F16 model is small

#1
by astrowonk - opened

I think there's boilerplate text when converting models leads to describing the F16 non-quantized models as "very large not recommend" but this model is a small SBERT model and even the F16 models is only ~50MB. Is there really a need to do quantization on models this small?

Sign up or log in to comment