VoxCPM / README.md
lihongjie
update
19d3e46
---
license: mit
language:
- en
- zh
base_model:
- VoxCPM
pipeline_tag: text-to-speech
library_name: transformers
tags:
- VoxCPM
- Speech
---
# VoxCPM
This version of VoxCPM has been converted to run on the Axera NPU using **w8a16** quantization.
Compatible with Pulsar2 version: 4.2
## Convert tools links:
For those who are interested in model conversion, you can try to export axmodel through the original repo :
[VoxCPM offical](https://github.com/OpenBMB/VoxCPM/)
[Pulsar2 Link, How to Convert LLM from Huggingface to axmodel](https://pulsar2-docs.readthedocs.io/en/latest/appendix/build_llm.html)
[AXera NPU HOST LLM Runtime](https://github.com/AXERA-TECH/VoxCPM.Axera)
## Support Platform
- AX650
- AX650N DEMO Board
- [M4N-Dock(爱芯派Pro)](https://wiki.sipeed.com/hardware/zh/maixIV/m4ndock/m4ndock.html)
- [M.2 Accelerator card](https://axcl-docs.readthedocs.io/zh-cn/latest/doc_guide_hardware.html)
## How to use
Download all files from this repository to the device
### 1. Run python demo
#### 1. Install voxcpm axinfer package
```
git clone -b 1.0.4-axmode_infer https://github.com/techshoww/VoxCPM.git
cd VoxCPM
pip3 install .
```
#### 2. Download zipenhancer
```
pip3 install modelscope
modelscope download --model iic/speech_zipenhancer_ans_multiloss_16k_base --local_dir iic/speech_zipenhancer_ans_multiloss_16k_base
```
#### 3. Run on Axera Device
Go to the root directory of this project. run:
```
python3 run_ax650.py
```
### 2. Run c++ demo
#### 1. Install transformers
```
pip3 install transformers>=4.56.2
```
#### 2. Start tokenizer server
```
python3 tokenizer.py --port 9999
```
#### 3. Run c++ demo
```
bash run_ax650.sh
```