| # MT-LLM |
|
|
| ><em> in International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2025 |
| <br> Authors: Lingwei Meng, Shujie Hu, Jiawen Kang, Yuejiao Wang, Wenxuan Wu, Xixin Wu, Xunying Liu, Helen Meng</em> |
|
|
| This repository contains the trained MT-LLM model for instruction-based multi-talker overlapped speech recognition. |
|
|
| Please check https://github.com/cuhealthybrains/MT-LLM |
|
|
| If you find our work is useful in your research, please cite the following paper: |
|
|
| ```bibtex |
| @inproceedings{meng2025mtllm, |
| title={Large Language Model Can Transcribe Speech in Multi-Talker Scenarios with Versatile Instructions}, |
| author={Meng, Lingwei and Hu, Shujie and Kang, Jiawen and Li, Zhaoqing and Wang, Yuejiao and Wu, Wenxuan and Wu, Xixin and Liu, Xunying and Meng, Helen}, |
| booktitle={ICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)}, |
| year={2025} |
| } |
| ``` |
|
|
| --- |
| license: mit |
| --- |
|
|