Meta's Llama-3 8B parameter base model trained on Alpaca dataset and outputed to 16bit GGUF Instruct model. Below is code for inference at command line using llama.cpp


./build/bin/main -m ./models/llama3_alpaca_dpo_GGUF-unsloth.F16.gguf \
-p '''Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.\n\n
### Instruction:\nWhy is the sky blue?\n\n
### Input:\n\n\n
### Response:\n'''

Downloads last month
5
GGUF
Model size
8B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support