Meta's Llama-3 8B parameter base model trained on Alpaca dataset and outputed to 16bit GGUF Instruct model. Below is code for inference at command line using llama.cpp
./build/bin/main -m ./models/llama3_alpaca_dpo_GGUF-unsloth.F16.gguf \
-p '''Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.\n\n
### Instruction:\nWhy is the sky blue?\n\n
### Input:\n\n\n
### Response:\n'''
- Downloads last month
- 5
Hardware compatibility
Log In to add your hardware
16-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support