FireRedTeam commited on
Commit
53925ef
·
verified ·
1 Parent(s): 374bb19

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,3 +1,19 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ## PUNC-BERT
2
+
3
+ This is a chinese-lert-base model finetuned for punctuation restoration, release mainly for [FireRedASR](https://github.com/FireRedTeam/FireRedASR) postprocessing.
4
+
5
+ Model restores the following punctuations -- [, 。 ? !]
6
+
7
+ ### License
8
+
9
+ Licensed under the Apache License, Version 2.0 (the "License");
10
+ you may not use this file except in compliance with the License.
11
+ You may obtain a copy of the License at
12
+
13
+ http://www.apache.org/licenses/LICENSE-2.0
14
+
15
+ Unless required by applicable law or agreed to in writing, software
16
+ distributed under the License is distributed on an "AS IS" BASIS,
17
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
18
+ See the License for the specific language governing permissions and
19
+ limitations under the License.
chinese-bert-wwm-ext_vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
chinese-lert-base-tokenizer/README.md ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - zh
4
+ license: "apache-2.0"
5
+ ---
6
+ ### LERT
7
+ LERT is a linguistically-motivated pre-trained language model.
8
+
9
+ Further information: https://github.com/ymcui/LERT/blob/main/README_EN.md
10
+
11
+ - **LERT: A Linguistically-motivated Pre-trained Language Model**
12
+ - *Yiming Cui, Wanxiang Che, Shijin Wang, Ting Liu*
13
+ - Paper link: https://arxiv.org/abs/2211.05344
chinese-lert-base-tokenizer/added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
chinese-lert-base-tokenizer/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": ".",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "directionality": "bidi",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "pooler_fc_size": 768,
21
+ "pooler_num_attention_heads": 12,
22
+ "pooler_num_fc_layers": 3,
23
+ "pooler_size_per_head": 128,
24
+ "pooler_type": "first_token_transform",
25
+ "position_embedding_type": "absolute",
26
+ "transformers_version": "4.16.2",
27
+ "type_vocab_size": 2,
28
+ "use_cache": true,
29
+ "vocab_size": 21128
30
+ }
chinese-lert-base-tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
chinese-lert-base-tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
chinese-lert-base-tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"init_inputs": []}
chinese-lert-base-tokenizer/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.pth.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22e94b720e3e53d0a113ff35cc60fbdb9bb78b92f55b895a246fb6a9ae6a9cae
3
+ size 1220496325
out_dict ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ <space> 0
2
+ , 1
3
+ 。 2
4
+ ? 3
5
+ ! 4