Inference with vLLM not working
#2
by llamameta - opened
OSError: openbmb/MiniCPM-V-2_6-gguf does not appear to have a file named
config.json. Checkout
'https://huggingface.co/openbmb/MiniCPM-V-2_6-gguf/tree/main' for available
files.
llamameta changed discussion title from Inference with vLLM to Inference with vLLM not working