tamarher's picture
Customize mlx-community model card
2d1fc12 verified
metadata
language:
  - zh
  - en
license: apache-2.0
library_name: mlx
pipeline_tag: text-to-speech
base_model: OpenMOSS-Team/MOSS-TTS-Local-Transformer
base_model_relation: quantized
tags:
  - mlx
  - tts
  - speech
  - voice-cloning
  - apple-silicon
  - quantized
  - 8bit

MOSS TTS Local Transformer — MLX 8-bit

This repository contains an MLX-native int8 conversion of MOSS TTS Local Transformer for single-speaker text-to-speech on Apple Silicon.

Note This repo is a community mirror of the canonical MLX conversion maintained by AppAutomaton at appautomaton/openmoss-tts-local-mlx.

Variants

Path Precision
mlx-int8/ int8 quantized weights

Model Details

How to Get Started

Command-line generation with mlx-speech:

Generate speech:

python scripts/generate_moss_local.py \
  --text "Hello, this is a test." \
  --output outputs/out.wav

Clone a voice:

python scripts/generate_moss_local.py \
  --mode clone \
  --text "This is a cloned voice." \
  --reference-audio reference.wav \
  --output outputs/clone.wav

Minimal Python usage:

from mlx_speech.generation import MossTTSLocalModel

model = MossTTSLocalModel.from_path("mlx-int8")

Notes

  • This repo contains the quantized MLX runtime artifact only.
  • The conversion keeps the original local TTS architecture and remaps weights explicitly for MLX inference.
  • The default runtime path uses W8Abf16 mixed precision with global and local KV cache enabled.
  • This mirror is a duplicated repo, not an automatically synchronized namespace mirror.

Links

License

Apache 2.0 — following the upstream license published with OpenMOSS-Team/MOSS-TTS-Local-Transformer.