Whisper Medium Korean (Zeroth fine-tune) โ€” MLX 4bit

ํ•œ๊ตญ์–ด ์Œ์„ฑ ์ธ์‹์„ ์œ„ํ•œ Whisper Medium fine-tune ๋ชจ๋ธ์„ Apple MLX ํ”„๋ ˆ์ž„์›Œํฌ์šฉ์œผ๋กœ 4bit ์–‘์žํ™”ํ•œ ๋ฒ„์ „์ž…๋‹ˆ๋‹ค.

์›๋ณธ: seastar105/whisper-medium-ko-zeroth (Whisper Medium์„ Zeroth Korean ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ fine-tune)

์š”์•ฝ

  • ๋ฒ ์ด์Šค: Whisper Medium (769M ํŒŒ๋ผ๋ฏธํ„ฐ)
  • Fine-tune: Zeroth Korean ASR corpus
  • ์–‘์žํ™”: 4bit (group size 64), mlx-examples/whisper/convert.py ์‚ฌ์šฉ
  • ๋””์Šคํฌ ํฌ๊ธฐ: 831 MB (์›๋ณธ fp16 2.8GB ๋Œ€๋น„ ์•ฝ 70% ๊ฐ์†Œ)
  • ์ถ”๋ก  RAM: ~1.26 GB
  • ํ”„๋ ˆ์ž„์›Œํฌ: Apple MLX (Apple Silicon ์ „์šฉ)

ํ•œ๊ตญ์–ด ์„ฑ๋Šฅ (Zeroth Korean test split)

์ง€ํ‘œ ๊ฐ’
CER 1.25%
WER 3.21%
RTF 0.055 (M3 16GB ๊ธฐ์ค€)

์›๋ณธ fp16 ๋ชจ๋ธ๊ณผ ๊ฑฐ์˜ ๋™์ผํ•œ ์ •ํ™•๋„๋ฅผ ์œ ์ง€ํ•˜๋ฉด์„œ ํฌ๊ธฐ์™€ ๋ฉ”๋ชจ๋ฆฌ๋ฅผ ํฌ๊ฒŒ ์ค„์˜€์Šต๋‹ˆ๋‹ค.

์‚ฌ์šฉ๋ฒ•

1) mlx-whisper ์ง์ ‘ ํ˜ธ์ถœ

pip install mlx-whisper
import mlx_whisper

result = mlx_whisper.transcribe(
    "audio.wav",
    path_or_hf_repo="youngouk/seastar-medium-ko-4bit-mlx",
    language="ko",
    word_timestamps=True,
)
print(result["text"])

2) meeting-transcriber ์•ฑ์—์„œ ์‚ฌ์šฉ

meeting-transcriber๋Š” ์ด ๋ชจ๋ธ์„ ๊ธฐ๋ณธ ์„ ํƒ์ง€๋กœ ์ œ๊ณตํ•˜๋Š” macOS ๋กœ์ปฌ ํšŒ์˜ ์ „์‚ฌ ์•ฑ์ž…๋‹ˆ๋‹ค.

์›น UI์—์„œ ์„ค์ • โ†’ ์Œ์„ฑ ์ธ์‹ ๋ชจ๋ธ (STT) โ†’ seastar medium-ko-zeroth (4bit)๋ฅผ ์„ ํƒํ•˜๋ฉด ์ž๋™ ๋‹ค์šด๋กœ๋“œ ๋ฐ ํ™œ์„ฑํ™”๋ฉ๋‹ˆ๋‹ค.

ํŒŒ์ผ ๊ตฌ์„ฑ

config.json              # MLX Whisper ๋ชจ๋ธ ์„ค์ • (์–‘์žํ™” ํŒŒ๋ผ๋ฏธํ„ฐ ํฌํ•จ)
weights.safetensors      # 4bit ์–‘์žํ™”๋œ ๊ฐ€์ค‘์น˜ (~415MB)

mlx-whisper ๋Ÿฐํƒ€์ž„์ด ์œ„ ๋‘ ํŒŒ์ผ์„ path_or_hf_repo= ์ธ์ž๋กœ ๋ฐ”๋กœ ๋กœ๋“œํ•ฉ๋‹ˆ๋‹ค. ํ† ํฌ๋‚˜์ด์ €๋Š” mlx-whisper๊ฐ€ ๋‚ด์žฅํ•œ multilingual vocab์„ ์‚ฌ์šฉํ•˜๋ฏ€๋กœ ๋ณ„๋„ ํŒŒ์ผ ๋ถˆํ•„์š”.

์–‘์žํ™” ํŒŒ๋ผ๋ฏธํ„ฐ

{
  "quantization": {
    "bits": 4,
    "group_size": 64
  }
}

์žฌํ˜„ ์ปค๋งจ๋“œ:

python mlx-examples/whisper/convert.py \
  --torch-name-or-path seastar105/whisper-medium-ko-zeroth \
  --mlx-path ./seastar-medium-ko-4bit \
  -q --q-bits 4 --q-group-size 64

๋ผ์ด์„ ์Šค

Apache License 2.0 โ€” ์›๋ณธ ๋ผ์ด์„ ์Šค๋ฅผ ๊ทธ๋Œ€๋กœ ์Šน๊ณ„ํ•ฉ๋‹ˆ๋‹ค.

์ œํ•œ ์‚ฌํ•ญ

  • Apple Silicon ์ „์šฉ: MLX ํ”„๋ ˆ์ž„์›Œํฌ๋Š” x86 CPU / CUDA์—์„œ ๋™์ž‘ํ•˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค. Intel Mac / Linux / Windows ์‚ฌ์šฉ์ž๋Š” ์›๋ณธ seastar105/whisper-medium-ko-zeroth๋ฅผ ์‚ฌ์šฉํ•˜์„ธ์š”.
  • ํ•œ๊ตญ์–ด ํŠนํ™”: Zeroth Korean ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ fine-tune๋˜์–ด ํ•œ๊ตญ์–ด ์™ธ ์–ธ์–ด ์„ฑ๋Šฅ์€ ๋ฒ ์ด์Šค Whisper Medium๋ณด๋‹ค ๋‚ฎ์„ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.
  • 4bit ์–‘์žํ™” ํŠน์„ฑ: ๋งค์šฐ ๋“œ๋ฌผ๊ฒŒ ํฌ๊ท€ ์–ดํœ˜์—์„œ ์›๋ณธ fp16๋ณด๋‹ค ์•ฝ๊ฐ„ ๋‚ฎ์€ ์ •ํ™•๋„๋ฅผ ๋ณด์ผ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค (์ธก์ •๋œ CER/WER ์ฐจ์ด๋Š” ๋ฌด์‹œ ๊ฐ€๋Šฅํ•œ ์ˆ˜์ค€).

์ถœ์ฒ˜ ยท ์ธ์šฉ

Downloads last month
62
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for youngouk/seastar-medium-ko-4bit-mlx

Finetuned
(1)
this model