metadata
library_name: llama.cpp
RickLLM
This is a GGUF format model uploaded using llama.cpp.
Model Details
- Model Format: GGUF (GPU/CPU inference using llama.cpp)
- Base Model: Unsloth
- Quantization: Q8_0
- Use Case: This model can be used with llama.cpp for efficient inference on both GPU and CPU.
Usage
This model can be used with llama.cpp. Example usage:
./main -m RickLLM.gguf -n 1024
License
Please refer to the original model's license for terms of use.