llm.create_chat_completion(
messages = "No input example has been defined for this model task."
)Unity Coder 30B - Q4_K_M + F16 GGUF
Quantized GGUF version of vishnuOI/unity-coder-30b for local inference.
- Base model: vishnuOI/unity-coder-30b
- Quantization: Q4_K_M + F16
- Format: GGUF (llama.cpp)
Usage
Load directly in LM Studio, Ollama, or llama.cpp.
Full Disclaimer:
Needs testing and tweaks- as it stands, the q4_k_m version with base LM Studio settings performs worse than some newer models without Unity Lora training- consider it more of an experiment. If someone has better experience with the f16 or has a good configuration, let me know!
- Downloads last month
- 208
Hardware compatibility
Log In to add your hardware
4-bit
16-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for wfwrigh/unity-coder-30b-GGUF
Base model
Qwen/Qwen3-Coder-30B-A3B-Instruct Adapter
vishnuOI/unity-coder-30b
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="wfwrigh/unity-coder-30b-GGUF", filename="", )