| | --- |
| | license: apache-2.0 |
| | language: |
| | - zh |
| | library_name: transformers |
| | tags: |
| | - Roberta |
| | - Chinese Pre-trained Language Model |
| | --- |
| | |
| | Please use 'XLMRoberta' related functions to load this model! |
| |
|
| | # MigBERT | 中文混合粒度预训练模型 |
| | [Character, Word, or Both? Revisiting the Segmentation Granularity for Chinese Pre-trained Language Models](https://arxiv.org/abs/2303.10893) |
| |
|
| | # Demo | 使用样例 |
| | https://github.com/xnliang98/MigBERT |
| |
|
| | # Citation |
| | 如果你觉得我们的工作对你有用,请在您的工作中引用我们的文章。 |
| |
|
| | If you find our resource or paper is useful, please consider including the following citation in your paper. |
| |
|
| | ``` |
| | @misc{liang2023character, |
| | title={Character, Word, or Both? Revisiting the Segmentation Granularity for Chinese Pre-trained Language Models}, |
| | author={Xinnian Liang and Zefan Zhou and Hui Huang and Shuangzhi Wu and Tong Xiao and Muyun Yang and Zhoujun Li and Chao Bian}, |
| | year={2023}, |
| | eprint={2303.10893}, |
| | archivePrefix={arXiv}, |
| | primaryClass={cs.CL} |
| | } |
| | ``` |