output = llm(
"Once upon a time,",
max_tokens=512,
echo=True
)
print(output)This is only in production mode i am trying to solve a error If you wanted to download this btw you can go to ollama and can use this locally. sorry for your inconvinence Ollama_link or can also use this locally.
ollama run hf.co/sk16er/vero-4b:Q4_K_M
Vero [Virtual Hero]
This is a light weight llm which you can use locally in your computer with
ollama run shushank/vero
But make sure you have ollama install! If not install from here - Ollama
To exit form this you can use this command
/bye
This is made by shushank
base model is gema by google
- Downloads last month
- 21
Hardware compatibility
Log In to add your hardware
4-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 1 Ask for provider support
Model tree for sk16er/vero-4b
Base model
google/gemma-3n-E4B
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="sk16er/vero-4b", filename="vero-4b-Q4_K_M.gguf", )