metadata
language: en
license: apache-2.0
tags:
- mlx
- fp16
- diffusion
- wedlm
library_name: mlx
pipeline_tag: text-generation
base_model: tencent/WeDLM-8B-Instruct
model_type: wedlm
WeDLM-8B-Instruct-MLX
This is a full-precision (fp16) MLX version of tencent/WeDLM-8B-Instruct for inference on Apple Silicon.
It currently does not work too well or provide meaningfull speedup due to lack of pre compilation. https://github.com/ZimengXiong/WeDLM-MLX/tree/main
Related Models
| Variant | HuggingFace |
|---|---|
| 4-bit | zimengxiong/WeDLM-8B-Instruct-MLX-4bit |
| 8-bit | zimengxiong/WeDLM-8B-Instruct-MLX-8bit |
| fp16 (this model) | zimengxiong/WeDLM-8B-Instruct-MLX |
License
This model inherits the license from the base model tencent/WeDLM-8B-Instruct.