Whisper Large-v3-Turbo Korean โ€” MLX 4bit

ํ•œ๊ตญ์–ด ์Œ์„ฑ ์ธ์‹์„ ์œ„ํ•œ Whisper Large-v3-Turbo fine-tune ๋ชจ๋ธ์„ Apple MLX ํ”„๋ ˆ์ž„์›Œํฌ์šฉ์œผ๋กœ 4bit ์–‘์žํ™”ํ•œ ๋ฒ„์ „์ž…๋‹ˆ๋‹ค.

์›๋ณธ: ghost613/whisper-large-v3-turbo-korean (Whisper Large-v3-Turbo๋ฅผ Zeroth Korean ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ fine-tune)

์š”์•ฝ

  • ๋ฒ ์ด์Šค: Whisper Large-v3-Turbo (809M ํŒŒ๋ผ๋ฏธํ„ฐ, 4 text layers)
  • Fine-tune: Zeroth Korean ASR corpus
  • ์–‘์žํ™”: 4bit (group size 64), mlx-examples/whisper/convert.py ์‚ฌ์šฉ
  • ๋””์Šคํฌ ํฌ๊ธฐ: 884 MB (์›๋ณธ fp16 3GB ๋Œ€๋น„ ์•ฝ 70% ๊ฐ์†Œ)
  • ์ถ”๋ก  RAM: ~1.31 GB
  • ํ”„๋ ˆ์ž„์›Œํฌ: Apple MLX (Apple Silicon ์ „์šฉ)

ํ•œ๊ตญ์–ด ์„ฑ๋Šฅ (Zeroth Korean test split)

์ง€ํ‘œ ๊ฐ’
CER 1.60%
WER 4.36%
RTF 0.056 (M3 16GB ๊ธฐ์ค€)

Large-v3-Turbo ์•„ํ‚คํ…์ฒ˜ ํŠน์„ฑ์ƒ seastar medium(4bit)๊ณผ ์œ ์‚ฌํ•œ ์ •ํ™•๋„๋ฅผ ์œ ์ง€ํ•˜๋ฉด์„œ ๋””์ฝ”๋”ฉ ๋ ˆ์ด์–ด๊ฐ€ ์–•์•„(4 layers) ์ฒซ ํ† ํฐ ์ง€์—ฐ์ด ์งง์Šต๋‹ˆ๋‹ค. ์†๋„ ์šฐ์„  ์‹œ๋‚˜๋ฆฌ์˜ค์— ์ ํ•ฉํ•ฉ๋‹ˆ๋‹ค.

์‚ฌ์šฉ๋ฒ•

1) mlx-whisper ์ง์ ‘ ํ˜ธ์ถœ

pip install mlx-whisper
import mlx_whisper

result = mlx_whisper.transcribe(
    "audio.wav",
    path_or_hf_repo="youngouk/ghost613-turbo-korean-4bit-mlx",
    language="ko",
    word_timestamps=True,
)
print(result["text"])

2) meeting-transcriber ์•ฑ์—์„œ ์‚ฌ์šฉ

meeting-transcriber๋Š” ์ด ๋ชจ๋ธ์„ "์†๋„ ์šฐ์„ " ์„ ํƒ์ง€๋กœ ์ œ๊ณตํ•˜๋Š” macOS ๋กœ์ปฌ ํšŒ์˜ ์ „์‚ฌ ์•ฑ์ž…๋‹ˆ๋‹ค.

์›น UI์—์„œ ์„ค์ • โ†’ ์Œ์„ฑ ์ธ์‹ ๋ชจ๋ธ (STT) โ†’ ghost613 turbo-korean (4bit)๋ฅผ ์„ ํƒํ•˜๋ฉด ์ž๋™ ๋‹ค์šด๋กœ๋“œ ๋ฐ ํ™œ์„ฑํ™”๋ฉ๋‹ˆ๋‹ค.

ํŒŒ์ผ ๊ตฌ์„ฑ

config.json              # MLX Whisper ๋ชจ๋ธ ์„ค์ • (์–‘์žํ™” ํŒŒ๋ผ๋ฏธํ„ฐ ํฌํ•จ)
weights.safetensors      # 4bit ์–‘์žํ™”๋œ ๊ฐ€์ค‘์น˜ (~440MB)

mlx-whisper ๋Ÿฐํƒ€์ž„์ด ์œ„ ๋‘ ํŒŒ์ผ์„ path_or_hf_repo= ์ธ์ž๋กœ ๋ฐ”๋กœ ๋กœ๋“œํ•ฉ๋‹ˆ๋‹ค. ํ† ํฌ๋‚˜์ด์ €๋Š” mlx-whisper๊ฐ€ ๋‚ด์žฅํ•œ multilingual vocab์„ ์‚ฌ์šฉํ•˜๋ฏ€๋กœ ๋ณ„๋„ ํŒŒ์ผ ๋ถˆํ•„์š”.

์–‘์žํ™” ํŒŒ๋ผ๋ฏธํ„ฐ

{
  "quantization": {
    "bits": 4,
    "group_size": 64
  }
}

์žฌํ˜„ ์ปค๋งจ๋“œ:

python mlx-examples/whisper/convert.py \
  --torch-name-or-path ghost613/whisper-large-v3-turbo-korean \
  --mlx-path ./ghost613-turbo-korean-4bit \
  -q --q-bits 4 --q-group-size 64

๋ผ์ด์„ ์Šค

Apache License 2.0 โ€” ์›๋ณธ ๋ผ์ด์„ ์Šค๋ฅผ ๊ทธ๋Œ€๋กœ ์Šน๊ณ„ํ•ฉ๋‹ˆ๋‹ค.

์ œํ•œ ์‚ฌํ•ญ

  • Apple Silicon ์ „์šฉ: MLX ํ”„๋ ˆ์ž„์›Œํฌ๋Š” x86 CPU / CUDA์—์„œ ๋™์ž‘ํ•˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค. Intel Mac / Linux / Windows ์‚ฌ์šฉ์ž๋Š” ์›๋ณธ ghost613/whisper-large-v3-turbo-korean๋ฅผ ์‚ฌ์šฉํ•˜์„ธ์š”.
  • ํ•œ๊ตญ์–ด ํŠนํ™”: Zeroth Korean ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ fine-tune๋˜์–ด ํ•œ๊ตญ์–ด ์™ธ ์–ธ์–ด ์„ฑ๋Šฅ์€ ๋ฒ ์ด์Šค Whisper Large-v3-Turbo๋ณด๋‹ค ๋‚ฎ์„ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.
  • 4bit ์–‘์žํ™” ํŠน์„ฑ: ๋งค์šฐ ๋“œ๋ฌผ๊ฒŒ ํฌ๊ท€ ์–ดํœ˜์—์„œ ์›๋ณธ fp16๋ณด๋‹ค ์•ฝ๊ฐ„ ๋‚ฎ์€ ์ •ํ™•๋„๋ฅผ ๋ณด์ผ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค (์ธก์ •๋œ CER/WER ์ฐจ์ด๋Š” ๋ฌด์‹œ ๊ฐ€๋Šฅํ•œ ์ˆ˜์ค€).

์ถœ์ฒ˜ ยท ์ธ์šฉ

Downloads last month
126
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for youngouk/ghost613-turbo-korean-4bit-mlx