| | --- |
| | library_name: transformers |
| | tags: [language-model, spoken-language, catastrophic-forgetting] |
| | --- |
| | |
| | # Model Description |
| |
|
| | This model is designed to address catastrophic forgetting in spoken language models during end-to-end training. It leverages innovative mitigation strategies to enhance model retention and performance over time. For more details, please refer to the paper by Hsiao et al. (2025) [here](https://arxiv.org/abs/2505.17496). |
| |
|
| | # Model Usage |
| |
|
| | To use this model, please check the [GitHub repository](https://github.com/chiyuanhsiao/ForgetSLM) for installation instructions, example code, and detailed usage guidelines. Ensure you have the required dependencies installed. |
| |
|
| |
|
| | # Citation |
| |
|
| | If you use this model in your research or applications, please cite it as follows: |
| |
|
| | **APA:** |
| | Hsiao, C.-Y., Lu, K.-H., Chang, K.-W., Yang, C.-K., Chen, W.-C., & Lee, H.-y. (2025). Analyzing Mitigation Strategies for Catastrophic Forgetting in End-to-End Training of Spoken Language Models. arXiv. https://arxiv.org/abs/2505.17496 |
| |
|
| | **BibTeX:** |
| | ``` |
| | @misc{hsiao2025analyzingmitigationstrategiescatastrophic, |
| | title={Analyzing Mitigation Strategies for Catastrophic Forgetting in End-to-End Training of Spoken Language Models}, |
| | author={Chi-Yuan Hsiao and Ke-Han Lu and Kai-Wei Chang and Chih-Kai Yang and Wei-Chih Chen and Hung-yi Lee}, |
| | year={2025}, |
| | eprint={2505.17496}, |
| | archivePrefix={arXiv}, |
| | primaryClass={cs.CL}, |
| | url={https://arxiv.org/abs/2505.17496}, |
| | } |
| | ``` |