--- tasks: - auto-speech-recognition domain: - audio model-type: - Classification frameworks: - onnx metrics: - f1_score license: apache-2.0 language: - cn tags: - FunASR - CT-Transformer - Alibaba - ICASSP 2020 widgets: - task: punctuation inputs: - type: text name: input title: 文本 examples: - name: 1 title: 示例1 inputs: - name: input data: 我们都是木头人不会讲话不会动 inferencespec: cpu: 1 #CPU数量 memory: 4096 --- # 模型介绍 ## Highlights 無量化,量化找官方的。 模型为[SenseVoice多语言语音理解模型Small](https://www.modelscope.cn/models/iic/SenseVoiceSmall)的onnx無量化导出版本,可以直接用来做生产部署,一键部署教程([点击此处](https://github.com/alibaba-damo-academy/FunASR/blob/main/runtime/readme_cn.md) ## [ModelScope-FunASR](https://github.com/alibaba-damo-academy/FunASR) [FunASR](https://github.com/alibaba-damo-academy/FunASR)提供可便捷本地或者云端服务器部署的离线文件转写服务,内核为FunASR已开源runtime-SDK。 集成了达摩院语音实验室在Modelscope社区开源的语音端点检测(VAD)、Paraformer-large语音识别(ASR)、标点恢复(PUNC) 等相关能力,拥有完整的语音识别链路,可以将几十个小时的音频或视频识别成带标点的文字,而且支持上百路请求同时进行转写。 [**最新动态**](https://github.com/alibaba-damo-academy/FunASR#whats-new) | [**环境安装**](https://github.com/alibaba-damo-academy/FunASR#installation) | [**介绍文档**](https://alibaba-damo-academy.github.io/FunASR/en/index.html) | [**服务部署**](https://github.com/alibaba-damo-academy/FunASR/blob/main/runtime/readme_cn.md) | [**模型库**](https://github.com/alibaba-damo-academy/FunASR/blob/main/docs/model_zoo/modelscope_models.md) | [**联系我们**](https://github.com/alibaba-damo-academy/FunASR#contact) ## 快速上手 ### docker安装 如果您已安装docker,忽略本步骤!! 通过下述命令在服务器上安装docker: ```shell curl -O https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/shell/install_docker.sh; sudo bash install_docker.sh ``` docker安装失败请参考 [Docker Installation](https://alibaba-damo-academy.github.io/FunASR/en/installation/docker.html) ### 镜像启动 通过下述命令拉取并启动FunASR runtime的docker镜像([获取最新镜像版本](https://github.com/alibaba-damo-academy/FunASR/blob/main/runtime/docs/SDK_advanced_guide_offline_zh.md)): ```shell sudo docker pull \ registry.cn-hangzhou.aliyuncs.com/funasr_repo/funasr:funasr-runtime-sdk-cpu-0.3.0 mkdir -p ./funasr-runtime-resources/models sudo docker run -p 10095:10095 -it --privileged=true \ -v $PWD/funasr-runtime-resources/models:/workspace/models \ registry.cn-hangzhou.aliyuncs.com/funasr_repo/funasr:funasr-runtime-sdk-cpu-0.3.0 ``` ### 服务端启动 docker启动之后,启动 funasr-wss-server服务程序: ```shell cd FunASR/runtime nohup bash run_server.sh \ --download-model-dir /workspace/models \ --vad-dir damo/speech_fsmn_vad_zh-cn-16k-common-onnx \ --model-dir damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-onnx \ --punc-dir damo/punc_ct-transformer_cn-en-common-vocab471067-large-onnx \ --lm-dir damo/speech_ngram_lm_zh-cn-ai-wesp-fst \ --itn-dir thuduj12/fst_itn_zh \ --hotword /workspace/models/hotwords.txt > log.out 2>&1 & ``` ### 客户端测试与使用 运行上面安装指令后,会在./funasr-runtime-resources(默认安装目录)中下载客户端测试工具目录samples([下载点击此处](https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/sample/funasr_samples.tar.gz)), 我们以Python语言客户端为例,进行说明,支持多种音频格式输入(.wav, .pcm, .mp3等),也支持视频输入(.mp4等),以及多文件列表wav.scp输入,其他版本客户端请参考文档([点击此处](https://alibaba-damo-academy.github.io/FunASR/en/runtime/docs/SDK_tutorial_zh.html#id5)) ```shell python3 wss_client_asr.py --host "127.0.0.1" --port 10095 --mode offline --audio_in "../audio/asr_example.wav" ``` 更详细用法介绍([点击此处](https://github.com/alibaba-damo-academy/FunASR/blob/main/runtime/docs/SDK_tutorial_zh.md)) ## 相关论文以及引用信息 ```BibTeX @inproceedings{chen2020controllable, title={Controllable Time-Delay Transformer for Real-Time Punctuation Prediction and Disfluency Detection}, author={Chen, Qian and Chen, Mengzhe and Li, Bo and Wang, Wen}, booktitle={ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)}, pages={8069--8073}, year={2020}, organization={IEEE} } ```