|
|
--- |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- en |
|
|
base_model: |
|
|
- prithivMLmods/Muscae-Qwen3-UI-Code-4B |
|
|
pipeline_tag: text-generation |
|
|
library_name: transformers |
|
|
tags: |
|
|
- code |
|
|
- gpt_oss |
|
|
- text-generation-inference |
|
|
- ui |
|
|
- web |
|
|
--- |
|
|
|
|
|
**Muscae-Qwen3-UI-Code-4B-GGUF** |
|
|
|
|
|
> Muscae-Qwen3-UI-Code-4B is a web-UI-focused model fine-tuned on UIGEN-T3-4B-Preview (built upon Qwen3-4B) for controlled Abliterated Reasoning and polished token probabilities, designed exclusively for experimental use. It excels at modern web UI coding tasks, structured component generation, and layout-aware reasoning, making it ideal for frontend developers, UI engineers, and research prototypes exploring structured code generation. |
|
|
|
|
|
## Model Files |
|
|
|
|
|
| File Name | Quant Type | File Size | |
|
|
| - | - | - | |
|
|
| Muscae-Qwen3-UI-Code-4B.BF16.gguf | BF16 | 8.05 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.F16.gguf | F16 | 8.05 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.F32.gguf | F32 | 16.1 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.Q2_K.gguf | Q2_K | 1.67 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.Q3_K_L.gguf | Q3_K_L | 2.24 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.Q3_K_M.gguf | Q3_K_M | 2.08 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.Q3_K_S.gguf | Q3_K_S | 1.89 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.Q4_K_M.gguf | Q4_K_M | 2.5 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.Q4_K_S.gguf | Q4_K_S | 2.38 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.Q5_K_M.gguf | Q5_K_M | 2.89 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.Q5_K_S.gguf | Q5_K_S | 2.82 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.Q6_K.gguf | Q6_K | 3.31 GB | |
|
|
| Muscae-Qwen3-UI-Code-4B.Q8_0.gguf | Q8_0 | 4.28 GB | |
|
|
|
|
|
## Quants Usage |
|
|
|
|
|
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) |
|
|
|
|
|
Here is a handy graph by ikawrakow comparing some lower-quality quant |
|
|
types (lower is better): |
|
|
|
|
|
 |