fix-json-GGUF / README.md
CronoBJS's picture
Update README.md
463b175 verified
metadata
license: apache-2.0
base_model: syntheticlab/fix-json
tags:
  - llama-cpp
  - gguf
  - lora
  - merged-model

fix-json (LoRA Merged GGUF)

Model creator: syntheticlab
Base model required: syntheticlab/fix-json (Meta-Llama-3.1-8B-Instruct)
GGUF conversion & quantization: by CronoBJS using llama.cpp

⚠️ Important:
This file is a merged LoRA adapter in GGUF format. It is not a standalone model.
You must also have the original base GGUF model (e.g., Meta-Llama-3.1-8B-Instruct.Q4_K_M.gguf) to run it.
Use the --lora or --lora-scaled flags in llama.cpp to apply it at runtime.


Special thanks

🙏 Thanks to Georgi Gerganov and the llama.cpp team for making all of this possible.


Running the Model

1️⃣ llama.cpp CLI

llama-cli -m "path/to/Meta-Llama-3.1-8B-Instruct.Q4_K_M.gguf" \
  --lora "path/to/syntheticlab-fix-json-lora.gguf" \
  -c 2048 -n 256 -p "Your prompt here"