bongsoo commited on
Commit
3b64fd9
·
1 Parent(s): db8e369

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -0
README.md CHANGED
@@ -34,6 +34,40 @@ model = AutoModel.from_pretrained('bongsoo/mbertV2.0')
34
  - loss : 훈련loss: 2.258400, 평가loss: 3.102096, [perplexity](https://github.com/kobongsoo/BERT/blob/master/bert/bert-perplexity-eval-V1.2.ipynb): 19.78158(bong_eval:1,500)
35
  - 훈련코드 [여기](https://github.com/kobongsoo/BERT/blob/master/bert/bert-MLM-Trainer-V1.2.ipynb) 참조
36
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
  ## Citing & Authors
38
 
39
  bongsoo
 
34
  - loss : 훈련loss: 2.258400, 평가loss: 3.102096, [perplexity](https://github.com/kobongsoo/BERT/blob/master/bert/bert-perplexity-eval-V1.2.ipynb): 19.78158(bong_eval:1,500)
35
  - 훈련코드 [여기](https://github.com/kobongsoo/BERT/blob/master/bert/bert-MLM-Trainer-V1.2.ipynb) 참조
36
 
37
+ ## Model Config
38
+ ```
39
+ {
40
+ "_name_or_path": "bert-base-multilingual-cased",
41
+ "architectures": [
42
+ "BertForMaskedLM"
43
+ ],
44
+ "attention_probs_dropout_prob": 0.1,
45
+ "classifier_dropout": null,
46
+ "directionality": "bidi",
47
+ "hidden_act": "gelu",
48
+ "hidden_dropout_prob": 0.1,
49
+ "hidden_size": 768,
50
+ "initializer_range": 0.02,
51
+ "intermediate_size": 3072,
52
+ "layer_norm_eps": 1e-12,
53
+ "max_position_embeddings": 512,
54
+ "model_type": "bert",
55
+ "num_attention_heads": 12,
56
+ "num_hidden_layers": 12,
57
+ "pad_token_id": 0,
58
+ "pooler_fc_size": 768,
59
+ "pooler_num_attention_heads": 12,
60
+ "pooler_num_fc_layers": 3,
61
+ "pooler_size_per_head": 128,
62
+ "pooler_type": "first_token_transform",
63
+ "position_embedding_type": "absolute",
64
+ "torch_dtype": "float32",
65
+ "transformers_version": "4.21.2",
66
+ "type_vocab_size": 2,
67
+ "use_cache": true,
68
+ "vocab_size": 152537
69
+ }
70
+ ```
71
  ## Citing & Authors
72
 
73
  bongsoo