Xinnian Liang commited on
Commit
e0f4d8c
·
1 Parent(s): ee238a2

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -0
README.md ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - zh
5
+ library_name: transformers
6
+ tags:
7
+ - Roberta
8
+ - Chinese Pre-trained Language Model
9
+ ---
10
+
11
+ Please use 'XLMRoberta' related functions to load this model!
12
+
13
+ # MigBERT | 中文混合粒度预训练模型 base 版本
14
+ [Character, Word, or Both? Revisiting the Segmentation Granularity for Chinese Pre-trained Language Models](https://arxiv.org/abs/2303.10893)
15
+
16
+ # Demo | 使用样例
17
+ https://github.com/xnliang98/MigBERT
18
+
19
+ # Citation
20
+ 如果你觉得我们的工作对你有用,请在您的工作中引用我们的文章。
21
+
22
+ If you find our resource or paper is useful, please consider including the following citation in your paper.
23
+
24
+ ```
25
+ @misc{liang2023character,
26
+ title={Character, Word, or Both? Revisiting the Segmentation Granularity for Chinese Pre-trained Language Models},
27
+ author={Xinnian Liang and Zefan Zhou and Hui Huang and Shuangzhi Wu and Tong Xiao and Muyun Yang and Zhoujun Li and Chao Bian},
28
+ year={2023},
29
+ eprint={2303.10893},
30
+ archivePrefix={arXiv},
31
+ primaryClass={cs.CL}
32
+ }
33
+ ```