Commit
·
840e994
1
Parent(s):
b8d770d
Upload 2 files
Browse files- latest +1 -0
- model_config.json +33 -0
latest
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
1
|
model_config.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_class": "GLMModel",
|
| 3 |
+
"tokenizer_type": "glm_ChineseSPTokenizer",
|
| 4 |
+
"num_layers": 48,
|
| 5 |
+
"hidden_size": 4096,
|
| 6 |
+
"num_attention_heads": 64,
|
| 7 |
+
"vocab_size": 50048,
|
| 8 |
+
"hidden_dropout": 0.1,
|
| 9 |
+
"attention_dropout": 0.1,
|
| 10 |
+
"layernorm_order": "pre",
|
| 11 |
+
"model_parallel_size": 1,
|
| 12 |
+
"max_sequence_length": 1025,
|
| 13 |
+
"block_lm": "true",
|
| 14 |
+
"masked_lm": false,
|
| 15 |
+
"bert_prob": 0.5,
|
| 16 |
+
"gpt_infill_prob": 0.5,
|
| 17 |
+
"gpt_min_ratio": 0.5,
|
| 18 |
+
"gap_sentence_prob": 0.0,
|
| 19 |
+
"gap_sentence_ratio": 0.15,
|
| 20 |
+
"avg_block_length": 3,
|
| 21 |
+
"short_seq_prob": 0.0,
|
| 22 |
+
"single_span_prob": 0.0,
|
| 23 |
+
"task_mask": "true",
|
| 24 |
+
"no_shuffle_block": false,
|
| 25 |
+
"no_block_position": false,
|
| 26 |
+
"sentinel_token": false,
|
| 27 |
+
"block_mask_prob": 0.0,
|
| 28 |
+
"context_mask_ratio": 0.0,
|
| 29 |
+
"random_position": false,
|
| 30 |
+
"cloze_eval": "true",
|
| 31 |
+
"old_checkpoint": false,
|
| 32 |
+
"tokenizer_model_type": "glm-10b"
|
| 33 |
+
}
|