lilfugu-transformers-8bit

8-bit quantized (bitsandbytes) version of lilfugu-transformers for CUDA / Linux. See the main model card for details.

2.2 GB — smallest CUDA variant.

Usage

pip install -U qwen-asr bitsandbytes
from qwen_asr import Qwen3ASRModel

model = Qwen3ASRModel.from_pretrained("holotherapper/lilfugu-transformers-8bit", device_map="auto")
result = model.transcribe("audio.wav", language="Japanese")
print(result[0].text)
Downloads last month
79
Safetensors
Model size
2B params
Tensor type
F32
·
F16
·
I8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for holotherapper/lilfugu-transformers-8bit

Quantized
(23)
this model