Xinnian Liang commited on
Commit ·
e0f4d8c
1
Parent(s): ee238a2
Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
language:
|
| 4 |
+
- zh
|
| 5 |
+
library_name: transformers
|
| 6 |
+
tags:
|
| 7 |
+
- Roberta
|
| 8 |
+
- Chinese Pre-trained Language Model
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
Please use 'XLMRoberta' related functions to load this model!
|
| 12 |
+
|
| 13 |
+
# MigBERT | 中文混合粒度预训练模型 base 版本
|
| 14 |
+
[Character, Word, or Both? Revisiting the Segmentation Granularity for Chinese Pre-trained Language Models](https://arxiv.org/abs/2303.10893)
|
| 15 |
+
|
| 16 |
+
# Demo | 使用样例
|
| 17 |
+
https://github.com/xnliang98/MigBERT
|
| 18 |
+
|
| 19 |
+
# Citation
|
| 20 |
+
如果你觉得我们的工作对你有用,请在您的工作中引用我们的文章。
|
| 21 |
+
|
| 22 |
+
If you find our resource or paper is useful, please consider including the following citation in your paper.
|
| 23 |
+
|
| 24 |
+
```
|
| 25 |
+
@misc{liang2023character,
|
| 26 |
+
title={Character, Word, or Both? Revisiting the Segmentation Granularity for Chinese Pre-trained Language Models},
|
| 27 |
+
author={Xinnian Liang and Zefan Zhou and Hui Huang and Shuangzhi Wu and Tong Xiao and Muyun Yang and Zhoujun Li and Chao Bian},
|
| 28 |
+
year={2023},
|
| 29 |
+
eprint={2303.10893},
|
| 30 |
+
archivePrefix={arXiv},
|
| 31 |
+
primaryClass={cs.CL}
|
| 32 |
+
}
|
| 33 |
+
```
|